Over the past few months, we've worked with many subreddits across different niches to implement our AI moderation bot with their community. Here, we highlight some stories from contrasting subreddits.
Dead by Daylight is a popular horror-style survival video game. Its community-run subreddit, r/deadbydaylight, serves as a hub for asking strategy questions, sharing gameplay experiences (and memes), and everything else related to Dead by Daylight.
Moderating the ~4,000 comments the subreddit receives each day can be a challenge though, given how heated conversations about gameplay can quickly grow toxic. So, turning to ModerateHatespeech’s SATI (Subreddit Anti-Toxicity Initiative), they’ve found great success in both cutting down moderator workload and finding previously unreported comments.
The challenge with r/deadbydaylight is the presence of game-specific terms, like “killer” example, which refers to a core character in the game. In the context of Dead by Daylight, certain comments that would be otherwise rule-breaking – like “kill yourself on hook and move on” (where “kill yourself on hook” is a strategy) – are actually harmless.
Working with r/deadbydaylight though, we’re able to balance and mitigate the effects of ambiguity on flagged comments, simultaneously ensuring a low rate of false positives and detection of genuine harassment. In fact, of the comments we flag to the moderation team at r/deadbydaylight, <10% are reported by other users. Because our bot scans through all the comments submitted to a subreddit, we catch many comments that moderators/vigilant user never discover, and instantly report them before they spread.
“I have been moderating r/DeadByDaylight for almost two years now, and one of the biggest issues I have seen is ensuring the comments of posts remained civil and courteous; people are not very good at reporting vitriolic comments, and so a lot of really uncalled-for stuff flew under the radar.
ToxicityModBot has been an absolute boon for this team: it has managed to catch so many comments that we are able to action on quickly (frequently before the user who got the hateful reply even notices), and has made our jobs so much easier in the process. I cannot recommend this bot enough for your sub.”
– moderator from /r/deadbydaylight
Reddit has a lot of politically oriented subreddits, but r/PoliticalDiscussion places a special emphasis on respectful, thought-provoking discourse. With chains of comments that span hundreds of words each, enforcing constructive, civil debates is extremely time-consuming, even if the sub only averages hundreds of comments a day. Plus, its well-known politics can be extremely polarizing!
By integrating our moderation system though, r/PoliticalDiscussion has been able to automatically remove rule-breaking content, seconds after its submitted (before users even see it) and identify more questionable content to moderators for further review.
The results have been nothing short of stellar – with just 4% of removed comments being false positives – and successfully detection of more than 80% of all the hateful, rule-breaking content submitted to the subreddit (based on retrospective analysis) – of which, only ~6% of was reported by other users. Naturally, this has translated into significant time savings for r/PoliticalDiscussion’s moderator team, totaling dozens of hours a month.
“ToxicityModBot has been very helpful in catching incivility preemptively, with the removal feature being quite strong with a low false positive rate when set to 95%. Reporting items between 90-95% has also been helpful. There's no perfect substitute in language processing for evaluating context, but in terms of time and effort saved, ToxicityModBot has been equal to an additional human moderator.”
– u/The_Egalitarian, moderator of r/PoliticalDiscussion