Toxic comments are associated with reduced activity of volunteer editors on Wikipedia (original) (raw)
Abstract
Wikipedia is one of the most successful collaborative projects in history. It is the largest encyclopedia ever created, with millions of users worldwide relying on it as the first source of information as well as for fact-checking and in-depth research. As Wikipedia relies solely on the efforts of its volunteer editors, its success might be particularly affected by toxic speech. In this paper, we analyze all 57 million comments made on user talk pages of 8.5 million editors across the six most active language editions of Wikipedia to study the potential impact of toxicity on editors’ behavior. We find that toxic comments are consistently associated with reduced activity of editors, equivalent to 0.5–2 active days per user in the short term. This translates to multiple human-years of lost productivity, considering the number of active contributors to Wikipedia. The effects of toxic comments are potentially even greater in the long term, as they are associated with a significantly increased risk of editors leaving the project altogether. Using an agent-based model, we demonstrate that toxicity attacks on Wikipedia have the potential to impede the progress of the entire project. Our results underscore the importance of mitigating toxic speech on collaborative platforms such as Wikipedia to ensure their continued success.
Significance Statement.
While the prevalence of toxic speech online is well studied, its true impact on the productivity of online communities remains largely unexplored. In this study, we focus on Wikipedia, which as the largest and most-read online reference, serves as a vital source of knowledge for millions of users worldwide. By analyzing all comments made over 20 years on user talk pages of 8.5 million editors across multiple language editions, we demonstrate that toxic speech is associated with a significant loss in the productivity of Wikipedia editors. These findings may have broad implications for large-scale collaborative projects and online communities, emphasizing the need to promote healthy and sustainable communication practices to protect crucial online information ecosystems and ensure their long-term success.
Introduction
Wikipedia is arguably one of the most successful collaborative projects in history. It has become the largest and most-read reference work ever created, and it is currently the fifth most popular website on the Internet (1). Millions of users worldwide rely on Wikipedia as their first source of information when encountering a new topic, for fact-checking and in-depth research (2). Even if caution might be required when consulting less actively maintained pages (3), numerous studies have shown that Wikipedia is a reliable source of information in areas ranging from political science (4) to pharmacology (5) and its accuracy is comparable to traditional encyclopedias (6) and textbooks (7).
One of the most remarkable aspects of Wikipedia’s success is that its content is exclusively created and curated by volunteer editors, known as Wikipedians. The English edition alone has more than 120,000 active editors (8). However, this volunteer-driven model also makes Wikipedia susceptible to the inherent challenges associated with maintaining such a large online community (9, 10). For example, it has been previously observed that Wikipedia is not free of conflict, particularly in the form of so-called edit wars (11), which impose significant costs on the project (12) and could negatively affect the quality of Wikipedia articles (13).
In this paper, we focus on the impact of toxic comments directed toward editors on their activity. This aspect is less studied, but potentially not less important, as affected by toxic comments, Wikipedians might reduce their contributions or abandon the project altogether, threatening the success of the platform (14).
Toxicity has been extensively studied on popular social media websites such as Twitter (15, 16), Reddit (17, 18), and similar platforms (19, 20). However, much of these research focuses on automated toxicity detection and prevalence estimation rather than on evaluating its impact (21). As an online encyclopedia, Wikipedia is often perceived as immune to toxicity and has a strict “No personal attacks” policy (22). Despite that, toxic speech and harassment have been previously observed on the platform (23–27). The effects of such behaviors on editors’ contributions are, however, not well understood nor well studied. The largest study to date relies on a voluntary opt-in survey of the 3,845 Wikipedians conducted in 2015 (24). It reports that 20% of users witnessing harassment have stopped contributing for a while, 17% considered not contributing anymore and 5% stopped contributing at all.
In this paper, we analyzed all 57 million comments made on user talk pages of editors on the six most active language editions of Wikipedia (English, German, French, Spanish, Italian, Russian) to understand the potential impact of toxic speech on editors’ contributions (see Methods and materials section for our definition of toxic comments). User talk pages are a place for editors to communicate with each other either on more personal topics or to extend their discussion from an article’s talk page. The majority of toxic comments are left on user talk pages (28). The comments we study were extracted from revision histories of talk pages and, thus, include even those toxic comments that were later archived or deleted by the page owner.
Figure 1 shows the activity of 50 randomly selected users who have received exactly one toxic comment. While some users are seemingly unaffected by a toxic comment, others temporarily reduce their activity or leave the project completely. The aim of our paper is to quantify this effect on the entire population of editors.
Fig. 1.
After receiving a toxic comment many users temporarily reduce their activity or leave the project completely. The figure shows the activity of 50 randomly selected users who received exactly one toxic comment. Blue squares indicate an active day, i.e. a day when at least one edit was done, starting from the first contribution of a given user. Red triangles correspond to toxic comments. Note that while some users are resilient and their activity is seemingly unaffected by toxic comments, many users temporarily reduce their activity or stop contributing altogether.
We estimate the number of lost active days associated with a toxic comment by comparing the number of active days before and after receiving a toxic comment. To account for potential baseline change, we have matched editors that received a toxic comment with similarly active editors who received a nontoxic comment. We have separately studied if toxic comments increase the probability of editors leaving the project altogether. Finally, we have used an agent-based model to model the potential impact of an increased number of toxic comments on Wikipedia.
Results
Loss of editor activity
To estimate the potential effect of a toxic comment, we compute the proportion of users who were active on day X before or after receiving a toxic comment (Fig. 2). We find that, on average, editors are more active near the time when they receive a toxic comment, with a peak at 24 h prior to the comment. At this time point, more than 40% of editors were active, as shown by the red line in Fig. 2a. This is a rather unsurprising observation since toxic comments are often made as a reaction to an edit made by a user and, thus, users are expected to be active around the time of a toxic comment. Note that if the timestamps around which the curve is centered are shuffled (black line in Fig. 2a) then this pattern disappears completely as expected.
Fig. 2.
After receiving a toxic comment, users become less active. On average, users are more active near the time when they receive a toxic comment (peak at zero for the red line in panel a). Average activity across all users who have received a toxic comment is lower in all 100 days after the event compared to the corresponding days before (dashed and solid red lines in panel b). This cannot be explained by a baseline drop in activity after a nontoxic comment (dashed and solid blue lines in panel b). Similar results hold not only for the English edition but also for the other five editions (c–g).
We also find that average activity across all users who have received a toxic comment is lower during all 100 days after the event compared to the corresponding days before (dashed and solid red lines in Fig. 2b), e.g. smaller number of users is active five days after receiving a toxic comment than five days before receiving it. To rule out the possibility that this is due to a general drop in activity over time or a drop in activity after any comment, we select a control group of users who have received a nontoxic comment, and whose average activity in the 100 days before the comment is the same as the average activity of users who received a toxic comment (see Methods and materials section for details).
We observe a similar characteristic peak around the nontoxic comment, likely due to both toxic and nontoxic comments being reactions to a contribution made by an editor. However, in contrast to a toxic comment, a nontoxic comment does not lead to a significant decrease in activity (dashed and solid blue lines in Fig. 2b). Similar results hold for all six language editions that we have examined (Fig. 2c–g).
We then estimate the lost activity associated with a toxic comment by computing the decrease in activity after a toxic comment, taking into account a potential baseline drop, i.e. by computing Δ=(Aftertoxic−Beforetoxic)−(Afternontoxic−Beforenontoxic). We find that this loss is statistically significant for all language editions studied (Table 1). We further explored the robustness of this result with respect to the toxicity threshold and potential filtering of users according to their activity. As expected, for higher toxicity thresholds, i.e. for more severely toxic comments, the effect is stronger (Supplementary Fig. S1). Considering only active users also leads to higher estimates; however, here we are reporting a conservative estimate, i.e. no filtering is used for results presented in Fig. 2 and Table 1.
Table 1.
Lost active days in the 100 days following a toxic comment.
Edition | Δ | _P_-value | Nusers |
---|---|---|---|
English | −1.207 | 2.6×10−66 | 36,332 |
German | −0.546 | 1.5×10−7 | 10,346 |
French | −1.851 | 4.8×10−9 | 2,239 |
Spanish | −0.563 | 8.6×10−3 | 2,446 |
Italian | −0.336 | 2.3×10−2 | 3,567 |
Russian | −1.219 | 7.8×10−4 | 1,134 |
While these results demonstrate that our findings are not limited to one language, they should not be used to compare effects between language editions, as there is no guarantee that the same toxicity threshold for the toxicity detection algorithm will have the same meaning in different languages.
Note that given that thousands of users have received at least one toxic comment (Supplementary Table S1), even a moderate loss per user could result in many human-years of lost productivity for Wikipedia in the short run. By multiplying the estimated loss per user from Table 1 by the number of users who have received at least one toxic comment, we could estimate the total loss of activity that is ranging from 5 human-years for Russian Wikipedia to 265 human-years for the English edition. The reason for the lasting effect of toxicity is that some new users might be discouraged by a toxic comment and choose to leave the project altogether after just a few contributions. This means that a single toxic comment could deprive Wikipedia of a potentially long-term contributor.
To further investigate this effect, we compare the probability of leaving Wikipedia after receiving a toxic comment with the probability of leaving Wikipedia after receiving a nontoxic comment.
Leaving Wikipedia
We observed that the probability of leaving Wikipedia after N contributions declines with N. PN(leaving) is approximately proportionate to N−α, where α ranges from 0.89 to 1.02, indicating a long-tailed distribution. While the probability of leaving the project after the first and only contribution is high (P1=47% for English Wikipedia), the risk of leaving Wikipedia drops to 0.7% for users who have made 100 contributions. To study the potential effects of toxic comments, we separately consider contributions that are followed by a toxic comment and contributions that are not followed by a toxic comment (see Methods and materials section for details). We find that the risk of an editor leaving after a toxic comment is consistently higher for all editions and regardless of the contribution number, see Fig. 3. We provide an analysis of the significance of these findings in Supplementary Fig. S4.
Fig. 3.
The probability of leaving Wikipedia after receiving a toxic comment is substantially higher than might be expected otherwise. For all six editions the probability of leaving declines with the number of contributions. At the same time, this probability is substantially higher after receiving a toxic comment than might be expected otherwise. Dots are probability estimates and solid lines are the best linear fit on a log-log scale.
Agent-based modeling
As has been demonstrated above, toxic comments increase the likelihood of editors abandoning Wikipedia. If enough editors leave, this could potentially impede the progress of the project as a whole. In order to estimate the potential impact of toxic comments, we model users’ behaviors by varying the toxicity of the environment, ranging from a nontoxic environment, where the probability of a user leaving follows the empirically observed nontoxic probability distribution, PNnon (blue dots in Fig. 3), to a highly toxic environment, where the probability of leaving corresponds to an empirically observed toxic probability distribution, PNtox (red dots in Fig. 3). We also consider a potential attack targeted at new users. In this scenario, each user receives a toxic comment after their first and second contributions, e.g. their probability of leaving after the first and second contribution is defined by PNtox, and after that follows the empirically observed PN.
For our modeling, we focus on a cohort of users who made their first contribution between the 4,000th and 6,000th day from the first recorded contribution to English Wikipedia in our dataset. We opted for this timeframe as it reflects Wikipedia’s current phase characterized by a relatively consistent number of active editors. This period follows the site’s initial exponential growth and a subsequent decline but comes before the anomalous increase in activity due to the COVID-19 pandemic (see Discussion section for details on these stages).
For our modeling, we employed an agent-based approach. Each day, agents (representing users) join Wikipedia and make their first contribution. The number of agents joining each day is equal to the actual count of first-time contributors to English Wikipedia on that particular day. After their first contribution, agents keep contributing, following a Poisson process, i.e. in such a way that the distance between two consecutive contributions, D, follows an exponential distribution: D∼Exp(λ), where λ is estimated from empirical data. After each contribution, the agent’s probability of leaving the project is determined by the toxicity level, T, and the empirically observed distributions PNnon and PNtox. In particular, after _N_’s contribution the user leaves the project with probability T*PNtox+(1−T)*PNnon. If the toxicity level is 0, then the probability of leaving follows the nontoxic distribution PNtox, and if the toxicity level is 1, then the probability of leaving follows the toxic distribution PNtox.
After the initial 2,000 days, no new agents join the project; however, we continue to model the behavior of the remaining agents for the subsequent 2,000 days, for which we have available empirical data for comparison.
Our model generally reproduces the dynamics of user activity (Fig. 4), though, as expected, it cannot account for a later COVID-19-induced spike in activity. We find that an extreme level of toxicity could effectively reduce the cohort to almost no users in the long run, compared to the sustained numbers in a nontoxic setting or as observed in the data. Additionally, targeted attacks on newcomers have the potential to significantly decrease the number of active users, posing a risk to the project. The detailed results of our modeling, showing the effects of different toxicity levels on user count, are presented in Supplementary Fig. S6.
Fig. 4.
High levels of toxicity and targeted attacks could significantly reduce the number of active editors. Modeling results for a cohort of editors making their first contribution during the relatively stable phase of Wikipedia (shaded region in the inset). The model reproduces the general dynamics of user activity (blue line) but, as expected, cannot capture the COVID-19-related spike in activity. An extreme level of toxicity (red line) could reduce the cohort to virtually no active users, contrasting with a nontoxic environment (green line) or actual activity (blue line). Targeted attacks on newcomers (orange line) have the potential to significantly reduce the number of active contributors.
Discussion
We conducted a large-scale analysis, covering all comments made on user talk pages of the six most active language editions of Wikipedia over a period of 20 years, and found that toxic comments are associated with a decreased activity of editors who have received these comments and an increased risk of them leaving the project altogether. Additionally, via agent-based modeling, we showed that toxicity attacks on Wikipedia have the potential to impede the progress of the entire project.
The main limitation of our study is its relatively narrow scope, as it focuses solely on the association between toxic comments left on user talk pages and the subsequent decrease in users’ activity. However, this approach allowed us to formulate our findings with precision and ensure their robustness. We believe that our study complements and extends existing studies on Wikipedia and online communities more broadly, and may serve as a foundation for further exploration of the effects of toxicity, as we discuss in this section.
Conflict on Wikipedia
Conflict on Wikipedia has already been a subject of numerous studies, with particular attention given to so-called “edit wars” (11, 29, 30). These arise when groups of editors, disagreeing about page content, repeatedly override each other’s contributions. It has been estimated that edit wars can impose substantial conflict and coordination costs on Wikipedia (12). Furthermore, it has been demonstrated that these costs increase over time and a smaller proportion of the total work by Wikipedians directly contributes to new article content. Conflict could also undermine content quality. For instance, the level of conflict on discussion pages, as assessed by raters, has been shown to negatively correlate with the quality of the corresponding Wikipedia articles (13).
In contrast to previous studies, our focus is on comments left on user talk pages rather than article talk pages. While this narrows the scope of our study, it also ensures that the comments we examine are directly addressed to a specific editor. Our approach also mitigates potential bias that could be introduced by the topic of an article. For instance, comments on talk pages linked to articles about violence might be misclassified as toxic by an algorithm due to the presence of highly negative keywords.
It is possible that toxic comments we observe on user talk pages are not independent from a broader conflict occurring elsewhere on Wikipedia. Therefore, it is conceivable that the effect we observe is not purely explained by toxic comments, but also by a broader conflict which leads both to a toxic comment on a user talk page and decreased activity of this user. Future research is needed to address this limitation and explore the context in which toxic comments occur.
It is worth noting, however, that it has already been established that toxicity on its own could lead users to stop contributing either temporarily or permanently, as this is what editors themselves report in surveys (24). Our study complements such studies by providing an estimate of the potential effects while also being performed on a scale that is not achievable by survey methods.
Stages of Wikipedia life cycle
Wikipedia has not grown linearly but has instead passed through several stages. It began with exponential growth (31), which subsequently slowed (32). Following that, the number of active users declined before Wikipedia entered its current stage, characterized by a relatively stable number of active users (33), with a slow decline observed in some language editions. A notable exception was a temporary spike in activity due to the COVID-19 pandemic (34). See Supplementary Fig. S5 for an illustration of these patterns in the editions studied in this paper.
It has been found that the main reason for halted growth is a sharp decline in the retention of newcomers (35). Specifically, with the project’s development, the rejection of newcomer contributions has increased, demotivating them and driving them away. Our results complement these findings by highlighting that newcomers are also particularly vulnerable to toxic comments. If users receive a toxic comment after their first or second contributions, their chances of continuing to contribute are 1.8 times lower compared to users who did not receive toxic comments.
Diversity of editors
Wikipedia is often considered a neutral and unbiased source of knowledge. In fact, this is ingrained in its “Neutral point of view” policy, which is officially one of the five fundamental principles of Wikipedia (36). However, the claim of neutrality should not be accepted uncritically (37). For instance, while Wikipedia mandates that its content is supported by reliable sources, the selection of these sources can significantly deviate from the norms of the expert knowledge community, introducing biases to Wikipedia content (38). Even if the content of articles is neutral, their coverage may be biased. It is well documented, for example, that biographies of women are underrepresented on Wikipedia (39). Wikipedia’s own rules might contribute to such biases. For instance, providing reliable sources as required by Wikipedia for biographies of women might be challenging because fewer sources exist on women due to historic inequalities (40). Another case in point is the Oral Citations project, which aimed to use oral citations for content on countries that are underrepresented in other sources (41). However, this initiative was met with opposition by the English Wikipedia community.
These content biases are closely connected to the lack of diversity among editors (38, 42). While estimates vary, the vast majority of Wikipedians are men (43). Notably, Wikipedia did not achieve its own goal of having at least 25% women editors by 2015 (44). This shortfall is a significant concern for the project, as diversity can improve the quality of content and reduce its biases (13, 45). While multiple barriers confront women editors on Wikipedia (40, 46, 47), toxicity is likely to be one of key factors contributing to the observed gender imbalance. Specifically, research has shown that while men and women are equally likely to face online harassment and abuse, women experience more severe violations (48). They are also more likely to be affected by such incidents and to self-censor in an attempt to prevent potential harassment (48). This has been confirmed in the Wikipedia context as well, where it has been demonstrated that the psychological experiences of women and men editors differ, leading to higher attrition rates among women (49). Similar results were found in another survey (24), showing that women experiencing toxicity are more likely to stop contributing in the future.
Overall, there are reasons to believe that toxicity might significantly undermine the diversity of Wikipedia editors, which can, in turn, compromise the quality of Wikipedia articles and introduce biases in its coverage. This underscores the importance of our findings. While most of the existing studies focus on the gender gap, we want to emphasize that the Wikipedia diversity problem goes beyond that, including racial, nonbinary, and other biases as well (50–52). For instance, we observed that many of the toxic comments in our data set include ethnic slurs. Future studies are needed to better understand the experiences of minority groups on Wikipedia and the effects that toxicity has on them.
Interventions
The Wikipedia community is well aware of the aforementioned problems, and there have been multiple efforts to address them through various interventions. Research into reward systems showed that while they might work effectively for already highly productive editors, they fail to motivate less active editors (53). Another study found no significant effect of positive rewards in online communities (54).
To address the gender gap in Wikipedia content, numerous events dedicated to creating entries about women were organized (46). An analysis of such interventions, which focused on two popular feminist interventions, confirmed that they succeeded in introducing content about women that would otherwise be missing (55). However, there is still a need to address the gender gap on a more systematic and sustainable level. For instance, one study showed that most of the women activists who attended editing workshops later chose not to continue contributing to Wikipedia, citing safety concerns as their primary reason (46). This issue was echoed in another study which identified safety as a core concern for women editors (56).
A suggested solution to this problem has been the red-flagging of harassment and harassers (46). However, the opinion that toxic comments are negligible and should be seen as merely over-enthusiastic participation is still present among editors (25). Furthermore, various anti-harassment measures have been declined multiple times by the community, as they were seen to slow the process of content creation (57, 58). Based on our findings, we believe there is a need to reevaluate these policies, and more research attention is required to understand the impact of potential interventions.
The wider role of peer-production systems
Wikipedia plays a crucial role in the global information infrastructure, aiming to provide millions of people with access to free, unbiased knowledge. Due to its reputation as a neutral and comprehensive information source, it has become a trusted first choice source of knowledge for many and its articles frequently appear in top search engine results (59, 60). In fact, studies have shown that Google search results rely heavily on Wikipedia, and the quality of these results significantly diminishes without Wikipedia (61). Beyond search engines, Wikipedia was shown to be valuable to other online communities such as Stack Exchange and Reddit (62).
While Wikipedia is arguably the most successful peer-production system, it is certainly not the only one. Others include hundreds of wikis hosted by Fandom, the numerous question-and-answer communities of Stack Exchange, and various other platforms ranging from online maps to online learning (33). Interestingly, for these projects, the same patterns that are typical of Wikipedia have been observed (63), i.e. the initial growth in number of contributors is followed by a decline characterized by a decreased retention of newcomers. This suggests that our findings might have broader implications for large-scale collaborative projects and online communities. It emphasizes the need to promote healthy and sustainable communication practices to protect crucial online information ecosystems and ensure their long-term success.
Methods and materials
Data and preprocessing
Comments on user talk pages
The Wikimedia Foundation provides publicly accessible dumps of all the different wikis’ content.a These dumps are updated on a regular basis, with complete revision history dumps generated once per month. For this paper, we used the English dump from 2021 November 1, the German dump from 2022 August 1, the French, Italian, and Spanish dumps from 2022 August 1, and the Russian dump from 2022 July 1. The data was obtained from a mirror hosted by the Umeå University, Sweden.b
From the dumps, the user talk pages were extracted. A user’s talk page is a place where other editors can communicate with the user either on more personal topics or to extend their discussion from an article talk page. When the comments left on the talk page are resolved or become too old, users can choose to archive them. This helps them keep better track of new incoming topics. Once archived, the old comments are not displayed on the talk page anymore but are rather linked in a subpage. Nevertheless, the entire history of the user talk page, as of any other page on Wikipedia, can be fully seen under the tab of revision history. The revision history records one entry for every edit made on the page saving each time the complete content of the page. Thus retrieving a single comment requires performing the difference between two consecutive revisions. The Wikimedia API does offer a method to compute the difference between two revisions, however, applying it on a scale that was necessary for this research was unfeasible. For that reason, we developed our own parser to extract comments as a difference between two versions of the page (64).
We excluded from our analysis talk pages that belong to unregistered users, e.g. users who are represented only by an IP address rather than a user name, because IP addresses are dynamic and it can not be assumed that one address represents a single user throughout Wikipedia history. Additionally, we have excluded comments made by officially registered bots. Comments that were made by users on their own pages are also not considered.
When extracting comments, we cleared wiki-specific formatting and HTML markup, i.e. removed links, attachments, or other formatting-specific sequences irrelevant to the actual content.
Contributions and active days
In order to extract information on users’ contributions, i.e. edits of Wikipedia pages made by them, we used the MediaWiki API to retrieve timestamps for each edit made by a given user. The resulting data set is publicly available in the project repository (64). The timestamps of contributions were then converted into active days. Specifically, each user i was represented as a binary vector ui=(ai1,ai2,…,aiN), where aid=1 if user i made at least one contribution, i.e. edited a Wikipedia page, within the 24-h period corresponding to day d and aid=0 otherwise. N is the number of days between the first recorded contribution in our data set and the last. The conversion from contribution count to active days was performed because it is hard to interpret and compare the total number of contributions between users as one large contribution could be equivalent to multiple smaller ones. Additionally, the size of a contribution does not necessarily reflect the effort put into it. While being active on a given day could still mean different levels of activity for different users, it represents a certain level of engagement with the project and is substantially different from not contributing at all on a given day.
Toxicity
The automatic detection of offensive language in online communities has been an active area of research since at least 2010 (65). Over the past decade, researchers have focused on detecting closely-related and intersecting types of offensive language such as toxicity, abusive language, and hate speech (66), see (67) for an overview of recent advancements in the field. In this paper, we use a model from the Perspective API (68) to identify toxic comments. This is a state-of-the-art toxicity detection algorithm that obtained competitive results at OffensEval-2019 competition (69) without any additional training on the contest data and is often used as a baseline system for toxicity detection (66). Perspective API is used across multiple platforms, including The New York Times, Der Spiegel, Le Monde, and El País. It uses BERT (Bidirectional Encoder Representations from Transformers) architecture (70) and is trained on comments from a variety of online sources, including Wikipedia. Each comment is labeled by 3–10 crowdsourced raters. Perspective models provide scores for several different attributes, see Supplementary Table S2 for the list of attributes and their definitions, see Supplementary Table S2 for examples of toxic comments, and see Supplementary Table S3 for the AUC (Area Under the Curve) scores for those languages and attributes that were used in this paper.
We define a toxic comment as a comment that has a score of at least 0.8 on any of the six dimensions provided by Perspective API. The 0.8 score means that on average 8 out of 10 raters would mark it as toxic. As this threshold can be considered arbitrary, we perform additional robustness checks using different toxicity thresholds. In particular, we compute activity loss not only for the threshold of 0.8 (Table 1) but for thresholds from 0.2 to 0.9. Additionally, we applied different activity filters, e.g. we separately compute an estimate only for those users who were active at least X days in the past 100 days where X varies from 0 to 50. This is done in order to ensure that the results are not exclusively driven by those users who had made few edits and then stopped contributing to the project. We perform this analysis for English Wikipedia as it is the largest edition. As shown in Supplementary Fig. S1, the estimate is typically in the range from −0.5 to −2 and significantly lower than zero for all activity thresholds and all toxicity thresholds higher than 0.3. Similarly, we have checked how the toxicity threshold affects the probability of leaving the project. As might be expected, results remain qualitatively the same for different toxicity thresholds but higher thresholds lead to more extreme results, e.g. the probability of leaving after a toxic comment with 0.9 score is even higher than after a toxic comment with toxicity score of 0.8 (Supplementary Fig. S3).
We also evaluated the robustness of our results with respect to misclassification errors. To achieve a realistic distribution of user activity, we repeatedly sampled 100,000 editors and their activity histories from the English Wikipedia data set. These sampled users were then divided into two groups: treatment and control. We investigated two distinct scenarios: one involving an equal split between the treatment and control groups and a second, more realistic, scenario where the treatment group constituted 1% of the control group.
In the treatment group, we randomly removed one active day from each user, thereby generating a true effect of one lost active day per user. We then introduced misclassification errors by generating false positives (moving users from control to treatment group) and false negatives (moving users from treatment to control group). Finally, we compared the estimated effect, as a function of the error rate, with the true effect.
We find that, generally, misclassification leads to the underestimation of the true effect, becoming more pronounced with higher error rates (Supplementary Fig. S2). The only exception is in the case of false negatives, i.e. undetected toxic comments, in the realistic scenario. Here, misclassification does not significantly bias the estimate, though it does increase its variance.
Perspective API accepts texts up to 20,480 bytes. As the majority of comments are well below this limit, we have excluded those that are larger.
Activity loss
Users who have received at least one toxic comment constitute our treatment group. For each user in this group, we select a random toxic comment they have received. We then center user activity around the timestamp, titox, of that toxic comment and convert the result to active days by calculating
sign(|{t∈Ti:t∈[titox+d*24*60*60,titox+(d+1)*24*60)}|),
where Ti is the set of timestamps of all contributions made by user i, and d is a day ranging from −100 to 100. Finally, the results are averaged over all users. We repeat the procedure of selecting a random toxic comment 100 times and report average results. However, since most users received only one toxic comment, there is little variation across simulations and the average over 100 simulations is almost identical to the result of a single simulation.
We then compare these results with a control group comprised of users who did not receive any toxic comments. However, a direct comparison is complicated because users who have received a toxic comment are, on average, more active than those who have not. This is probably due to the fact that each contribution could lead to a toxic response with a certain probability. Hence, the more contributions a user makes, the higher the likelihood of receiving a toxic comment and thereby being in the treatment group.
Specifically, if each contribution can lead to a toxic comment with a probability p, then the probability of receiving at least one toxic comment depends on the number of contribution, N: P(gettoxiccomment)=1−(1−p)N(1).
To ensure our control group is similarly active as the treatment group, we randomly select users with a probability based on the number of their contributions using formula (1). Users selected in this manner form the control group. For these users, we then pick a nontoxic comment at random, center their activity around its timestamp, and follow the same procedure used for the treatment group.
To test for the significance of the results, we compute 95% bootstrapped confidence intervals for each estimate.
Probability of leaving
For each toxic comment, we find the closest in time contribution that precedes that comment. We define such contributions as “contributions followed by a toxic comment” and compare the probability of leaving after such contributions with the probability of leaving after other contributions. The probability of leaving after N contributions is estimated as a fraction of users who have made exactly N contributions among users who have made at least N contributions. As the probability of leaving strongly depends on N, we make a comparison separately for each contribution number N∈[1,100]. For N>100 the number of users is too small to provide reliable estimates for comparison.
Supplementary Material
pgad385_Supplementary_Data
Acknowledgments
We acknowledge the Master’s thesis by Brückner (71), which identified a potential pattern in data and provided an inspiration for the design of the study presented in this paper. The initial data collection and experiments were carried out as part of Camelia Oprea’s Master’s thesis (72). We thank Liubov Tupikina and David Garcia for their valuable discussions regarding the results presented in this article. We thank the anonymous reviewers for their insightful comments and suggestions.
Footnotes
Contributor Information
Ivan Smirnov, Graduate Research School, University of Technology Sydney, 15 Broadway, Sydney 2007, Australia.
Camelia Oprea, Department of Computer Science, RWTH Aachen University, Ahornstrasse 55, Aachen 52074, Germany.
Markus Strohmaier, Business School, University of Mannheim, L 15 1–6, Mannheim 68161, Germany; GESIS—Leibniz Institute for the Social Sciences, Unter Sachsenhausen 6–8, Köln 50667, Germany; Complexity Science Hub Vienna, Josefstaedter Strasse 39, Vienna 1080, Austria.
Supplementary Material
Supplementary material is available at PNAS Nexus online.
Funding
The publication of this article was funded by the University of Mannheim.
Author Contributions
I.S., C.O., and M.S. designed the study; I.S. and C.O. collected and analyzed the data; I.S., C.O., and M.S. wrote the manuscript; I.S. revised the manuscript.
Previous Presentation
These results were previously presented at International Conference on Computational Social Science 2023.
Preprints
A preprint of this article is published at https://doi.org/10.48550/arXiv.2304.13568
Data Availability
The data underlying this article is available in Open Science Framework at https://osf.io/2qyxj/.
References
- 1.Semruch. 2023. Most visited websites in the world [accessed 2023 Sept]. https://www.semrush.com/website/top/.
- 2.Singer P, et al. 2017. Why we read Wikipedia. In: Proceedings of the 26th International Conference on World Wide Web. Perth, Australia: Association for Computing Machinery. p. 1591–1600.
- 3.Bruckman AS. 2022. Should you believe Wikipedia? Online communities and the construction of knowledge. Cambridge, UK: Cambridge University Press. [Google Scholar]
- 4.Brown AR. 2011. Wikipedia as a data source for political scientists: accuracy and completeness of coverage. PS: Political Sci Politics. 44(2):339–343. [Google Scholar]
- 5.Clauson KA, Polen HH, Boulos MNK, Dzenowagis JH. 2008. Scope, completeness, and accuracy of drug information in Wikipedia. Ann Pharmacother. 42(12):1814–1821. [DOI] [PubMed] [Google Scholar]
- 6.Giles J. 2005. Internet encyclopaedias go head to head. Nature. 438(15):900–901. [DOI] [PubMed] [Google Scholar]
- 7.Kräenbring J, et al. 2014. Accuracy and completeness of drug information in Wikipedia: a comparison with standard textbooks of pharmacology. PLoS ONE. 9(9):e106930. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Wikipedia. 2023. List of Wikipedias [accessed 2023 Sept]. https://en.wikipedia.org/wiki/List˙of˙Wikipedias.
- 9.Kraut RE, Resnick P. 2012. Building successful online communities: evidence-based social design. Cambridge, MA, USA: MIT Press. [Google Scholar]
- 10.Keegan B, Fiesler C. 2017. The evolution and consequences of peer producing Wikipedia’s rules. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 11. Montreal, Canada: Association for the Advancement of Artificial Intelligence. p. 112–121.
- 11.Yasseri T, Sumi R, Rung A, Kornai A, Kertész J. 2012. Dynamics of conflicts in Wikipedia. PLoS ONE. 7(6):e38869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Kittur A, Suh B, Pendleton BA, Chi EH. 2007. He says, she says: conflict and coordination in Wikipedia. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. San Jose, CA, USA: Association for Computing Machinery. p. 453–462.
- 13.Arazy O, Nov O, Patterson R, Yeo L. 2011. Information quality in Wikipedia: the effects of group composition and task conflict. J Manag Inf Syst. 27(4):71–98. [Google Scholar]
- 14.Preece J. 2001. Sociability and usability in online communities: determining and measuring success. Behav Inf Technol. 20(5):347–356. [Google Scholar]
- 15.Chatzakou D, et al. 2017. Measuring# gamergate: a tale of hate, sexism, and bullying. In: Proceedings of the 26th International Conference on World Wide Web. Perth, Australia: Association for Computing Machinery. p. 1285–1290.
- 16.Guberman J, Schmitz C, Hemphill L. 2016. Quantifying toxicity and verbal violence on twitter. In: Proceedings of the 19th ACM Conference on Computer Supported Cooperative Work and Social Computing. San Francisco, CA, USA: Association for Computing Machinery. p. 277–280.
- 17.Xia Y, Zhu H, Lu T, Zhang P, Gu N. 2020. Exploring antecedents and consequences of toxicity in online discussions: a case study on Reddit. Proc ACM Hum-Comput Interact. 4:1–23. [Google Scholar]
- 18.Almerekhi H, Jansen BJ, Kwak H. 2020. Investigating toxicity across multiple Reddit communities, users, and moderators. In: Companion Proceedings of the Web Conference. Taipei, Taiwan: Association for Computing Machinery. p. 294–298.
- 19.Wich M, et al. 2022. Introducing an abusive language classification framework for telegram to investigate the german hater community. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 16. Atlanta, GA, USA: Association for the Advancement of Artificial Intelligence. p. 1133–1144.
- 20.Silva L, Mondal M, Correa D, Benevenuto F, Weber I. 2016. Analyzing the targets of hate in online social media. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 10. Cologne, Germany: Association for the Advancement of Artificial Intelligence. p. 687–690.
- 21.Kiritchenko S, Nejadgholi I, Fraser KC. 2021. Confronting abusive language online: a survey from the ethical and human rights perspective. J Artif Intell Res. 71:431–478. [Google Scholar]
- 22.Wikipedia . Personal attacks, 2023. https://en.wikipedia.org/wiki/Wikipedia:No_personal_attacks.
- 23.Arazy O, Yeo L, Nov O. 2013. Stay on the Wikipedia task: when task-related disagreements slip into personal and procedural conflicts. J Am Soc Inf Sci Technol. 64(8):1634–1648. [Google Scholar]
- 24.Wikimedia Foundation . Harassment survey, 2023. https://upload.wikimedia.org/wikipedia/commons/5/52/Harassment˙Survey˙2015˙-˙Results˙Report.pdf.
- 25.Corple DJ. 2016. Beyond the gender gap: understanding women's participation in Wikipedia. Open Access Theses. 936. https://docs.lib.purdue.edu/open_access_theses/936.
- 26.Wulczyn E, Thain N, Dixon L. 2017. Ex machina: personal attacks seen at scale. In: Proceedings of the 26th International Conference on World Wide Web. Perth, Australia: Association for Computing Machinery. p. 1391–1399.
- 27.Raish M. 2019. Identifying and classifying harassment in arabic Wikipedia: a “netnography” [accessed 2023 Sept]. https://upload.wikimedia.org/wikipedia/commons/7/78/Arabic_Harassment_Netnography_Report.pdf.
- 28.Qu I, Thain N, Hua Y. 2019. Wikidetox visualization. In: Wiki Workshop; San Francisco, CA, USA.
- 29.Sumi R, Yasseri T. 2011. Edit wars in Wikipedia. In: 2011 IEEE Third International Conference on Privacy, Security, Risk and Trust and 2011 IEEE Third International Conference on Social Computing. Boston, MA, USA: Institute of Electrical and Electronics Engineers. p. 724–727.
- 30.Chhabra A, Kaur R, Iyengar SRS. 2020. Dynamics of edit war sequences in Wikipedia. In: Proceedings of the 16th International Symposium on Open Collaboration. Virtual Event, Spain: Association for Computing Machinery. p. 1–10.
- 31.Almeida RB, Mozafari B, Cho J. 2007. On the evolution of Wikipedia. In: Proceedings of the International Conference on Web and Social Media (ICWSM). Boulder, CO, USA: Association for the Advancement of Artificial Intelligence.
- 32.Suh B, Convertino G, Chi EH, Pirolli P. 2009. The singularity is not near: slowing growth of Wikipedia. In: Proceedings of the 5th International Symposium on Wikis and Open Collaboration. Orlando, FL, USA: Association for Computing Machinery. p. 1–10.
- 33.Hill BM, Shaw A. 2020. Wikipedia and the end of open collaboration. Wikipedia, 20.
- 34.Ruprechter T, et al. 2021. Volunteer contributions to Wikipedia increased during Covid-19 mobility restrictions. Sci Rep. 11(1):21505. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Halfaker A, Geiger RS, Morgan JT, Riedl J. 2013. The rise and decline of an open collaboration system: how Wikipedia’s reaction to popularity is causing its decline. Am Behav Sci. 57(5):664–688. [Google Scholar]
- 36.Wikipedia: neutral point of view [accessed 2023 Oct]. https://en.wikipedia.org/wiki/Wikipedia:Neutral˙point˙of˙view.
- 37.Matei SA, Dobrescu C. 2011. Wikipedia’s “neutral point of view”: settling conflict through ambiguity. Inf Soc. 27(1):40–51. [Google Scholar]
- 38.Luyt B. 2012. The inclusivity of Wikipedia and the drawing of expert boundaries: an examination of talk pages and reference lists. J Am Soc Inf Sci Technol. 63(9):1868–1878. [Google Scholar]
- 39.Wagner C, Garcia D, Jadidi M, Strohmaier M. 2015. It’s a man’s Wikipedia? Assessing gender inequality in an online encyclopedia. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 9. Oxford, UK: Association for the Advancement of Artificial Intelligence. p. 454–463.
- 40.Ford H, Wajcman J. 2017. ‘anyone can edit’, not everyone does: Wikipedia’s infrastructure and the gender gap. Soc Stud Sci. 47(4):511–527. [DOI] [PubMed] [Google Scholar]
- 41.Gallert P, Van der Velden M. 2013. Reliable sources for indigenous knowledge: dissecting Wikipedia’s catch-22 [accessed 2023 Sept]. https://upload.wikimedia.org/wikipedia/commons/5/51/Indigenous_Knowledge_for_Wikipedia.pdf.
- 42.Lam STK, et al. 2011. Wp: clubhouse? an exploration of Wikipedia’s gender imbalance. In: Proceedings of the 7th International Symposium on Wikis and Open Collaboration. Mountain View, CA, USA: Association for Computing Machinery. p. 1–10.
- 43.Hill BM, Shaw A. 2013. The Wikipedia gender gap revisited: characterizing survey response bias with propensity score estimation. PLoS ONE. 8(6):e65782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Wikipedia ‘completely failed’ to fix gender imbalance [accessed 2023 Oct]. https://www.bbc.com/news/av/business-28701772.
- 45.Sydow M, Baraniak K, Teisseyre P. 2017. Diversity of editors and teams versus quality of cooperative work: experiments on Wikipedia. J Intell Inf Syst. 48:601–632. [Google Scholar]
- 46.Lir SA. 2021. Strangers in a seemingly open-to-all Website: the gender bias in Wikipedia. Equal Divers Incl Int J. 40(7):801–818. [Google Scholar]
- 47.Menking A, Erickson I. 2015. The heart work of Wikipedia: gendered, emotional labor in the world’s largest online encyclopedia. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. Seoul, Republic of Korea: Association for Computing Machinery. p. 207–210.
- 48.Lenhart A, Ybarra M, Zickuhr K, Prive-Feeney M. 2016. Online harassment, digital abuse, and cyberstalking in America [accessed 2023 Sept]. https://www.datasociety.net/pubs/oh/Online_Harassment_2016.pdf.
- 49.Bear JB, Collier B. 2016. Where are the women in Wikipedia? understanding the different psychological experiences of men and women in Wikipedia. Sex Roles. 74:254–265. [Google Scholar]
- 50.Lemieux ME, Zhang R, Tripodi F. 2023. “too soon” to count? How gender and race cloud notability considerations on Wikipedia. Big Data Soc. 10(1):20539517231165490. [Google Scholar]
- 51.Tripodi F. 2023. Ms. categorized: gender, notability, and inequality on Wikipedia. New Media Soc. 25(7):1687–1707. [Google Scholar]
- 52.Field A, Park CY, Lin KZ, Tsvetkov Y. 2022. Controlled analyses of social biases in Wikipedia bios. In: Proceedings of the ACM Web Conference 2022. Lyon, France: Association for Computing Machinery. p. 2624–2635.
- 53.Restivo M, van de Rijt A. 2014. No praise without effort: experimental evidence on how rewards affect Wikipedia’s contributor community. Inf Commun Soc. 17(4):451–462. [Google Scholar]
- 54.Cheng J, Danescu-Niculescu-Mizil C, Leskovec J. 2014. How community feedback shapes user behavior. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 8. Ann Arbor, MI, USA: Association for the Advancement of Artificial Intelligence. p. 41–50. [DOI] [PMC free article] [PubMed]
- 55.Langrock I, González-Bailón S. 2022. The gender divide in Wikipedia: quantifying and assessing the impact of two feminist interventions. J Commun. 72(3):297–321. [Google Scholar]
- 56.Menking A, Erickson I, Pratt W. 2019. People who can take it: how women Wikipedians negotiate and navigate safety. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. Glasgow, UK: Association for Computing Machinery. p. 1–14.
- 57.Wikimedia Foundation . Community wishlist survey. Bots and gadgets, 2023. https://meta.wikimedia.org/wiki/Community˙Wishlist˙Survey˙2015/Bots˙and˙gadgets.
- 58.Wikimedia Foundation . Community wishlist survey. Anti-harassment, 2023. https://meta.wikimedia.org/wiki/Community˙Wishlist˙Survey˙2021/Anti-harassment.
- 59.Vincent N, Hecht B. 2021. A deeper investigation of the importance of Wikipedia links to search engine results. Proc ACM Hum-Comput Interact. 5(CSCW1):1–15.36644216 [Google Scholar]
- 60.Vincent N, Johnson I, Sheehan P, Hecht B. 2019. Measuring the importance of user-generated content to search engines. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 13. Munich, Germany: Association for the Advancement of Artificial Intelligence. p. 505–516.
- 61.McMahon C, Johnson I, Hecht B. 2017. The substantial interdependence of Wikipedia and Google: a case study on the relationship between peer production communities and information technologies. In: Proceedings of the International AAAI Conference on Web and Social Media. Vol. 11. Montreal, Canada: Association for the Advancement of Artificial Intelligence. p. 142–151.
- 62.Vincent N, Johnson I, Hecht B. 2018. Examining Wikipedia with a broader lens: quantifying the value of Wikipedia’s relationships with other large-scale online communities. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. Montreal, Canada: Association for Computing Machinery. p. 1–13.
- 63.TeBlunthuis N, Shaw A, Hill BM. 2018. Revisiting “the rise and decline” in a population of peer production projects. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. Montreal, Canada: Association for Computing Machinery. p. 1–7.
- 64.Smirnov I, Oprea C. 2023. Toxic comments reduce activity of volunteer editors on Wikipedia: data & code, 2023. https://osf.io/2qyxj. [DOI] [PMC free article] [PubMed]
- 65.Xu Z, Zhu S. 2010. Filtering offensive language in online communities using grammatical relations. In: Proceedings of the Seventh Annual Collaboration, Electronic Messaging, Anti-Abuse and Spam Conference. Redmond, WA, USA: Association for Computing Machinery. p. 1–10.
- 66.Fortuna P, Soler J, Wanner L. 2020. Toxic, hateful, offensive or abusive? what are we really classifying? an empirical analysis of hate speech datasets. In: Proceedings of the 12th Language Resources and Evaluation Conference. Marseille, France: Association for Computational Linguistics. p. 6786–6794.
- 67.Zampieri M, et al. 2020. Semeval-2020 task 12: multilingual offensive language identification in social media (offenseval 2020), arXiv, arXiv:2006.07235, preprint: not peer reviewed.
- 68.Perspective API . Technical documentation, 2023. https://developers.perspectiveapi.com/s/about-the-api.
- 69.Zampieri M, et al. 2019. Semeval-2019 task 6: identifying and categorizing offensive language in social media (offenseval), arXiv, arXiv:1903.08983, preprint: not peer reviewed.
- 70.Lees A, et al. 2022. A new generation of perspective API: efficient multilingual character-level transformers, arXiv, arXiv:2202.11176, preprint: not peer reviewed.
- 71.Brückner S. 2021. Modeling sociodemographic attributes of Wikipedia editors [master thesis]. RWTH Aachen.
- 72.Oprea C. 2022. Determining the impact of toxicity on Wikipedia’s talk pages [master thesis]. RWTH Aachen.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
pgad385_Supplementary_Data
Data Availability Statement
The data underlying this article is available in Open Science Framework at https://osf.io/2qyxj/.