Content Moderation Case Study: Decentralized Social Media Platform Mastodon Deals With An Influx Of Gab Users (2019)
from the decentralized-content-moderation-challenges dept
Summary: Formed as a more decentralized alternative to Twitter that allowed users to more directly moderate the content they wanted to see, Mastodon has experienced slow, but steady, growth since its inception in 2016.
Unlike other social media networks, Mastodon is built on open-source software and each "instance" (server node) of the network is operated by users. These separate "instances" can be connected with others via Mastodon's interlinked "fediverse." Or they can remain independent, creating a completely siloed version of Mastodon that has no connection with the service's larger "fediverse."
This puts a lot of power in the hands of the individuals who operate each instance: they can set their own rules, moderate content directly, and prevent anything the "instance" and its users find undesirable from appearing on their servers. But the larger "fediverse" -- with its combined user base -- poses moderation problems that can't be handled as easily as those presenting themselves on independent "instances." The connected "fediverse" allows instances to interact with each other, allowing unwanted content to appear on servers that are trying to steer clear of it.
That's where Gab -- another Twitter alternative -- enters the picture. Gab has purposely courted users banned from other social media services. Consequently, the platform has developed a reputation for being a haven for hate speech, racists, and bigots of all varieties. This toxic collection of content/users led to both Apple and Google banning Gab's app from their app stores.
Faced with this app ban, Gab began looking for options. It decided to create its own Mastodon instance. With its server now technically available to everyone in the Mastodon "fediverse," those not explicitly blocking Gab's "instance" could find Gab content available to its users -- and also allow for Gab’s users to direct content to their own users. It also allowed Gab to utilize the many different existing Mastodon apps to sidestep the app bans handed down by Google and Apple.
Decisions to be made by Mastodon:
-
Should Gab (and its users) be banned from setting up "instances," given that they likely violate the Mastodon Server Covenant?
-
Is it possible to moderate content across a large number of independent nodes?
-
Is this even an issue for Mastodon itself to deal with, given that the individuals running different servers can decide for themselves whether or not to allow federation with the Gab instance?
-
Given the open source and federated nature of Mastodon, would there reasonably be any way to stop Gab from using Mastodon?
-
Will moderation efforts targeting the "fediverse" undercut the independence granted to "instance" owners?
-
Do attempts to attract more users create moderation friction when the newly-arriving users create content Mastodon was created to avoid?
-
If Mastodon continues to scale, will it always face challenges as certain instances are created to appeal to audiences that the rest of the “fediverse” is trying to avoid?
-
Can a federated system, in which unique instances choose not to federate with another instance, such as Gab, work as a form of “moderation-by-exclusion”?
On the other hand, most "fediverse" members would be unlikely to have to deal with Gab or its users, considering the content contained in Gab's "instance" routinely violates the Mastodon "covenant." Violating these rules prevents instances from being listed by Mastodon itself, lowering the chances of other "instance" owners inadvertently adding toxic content and users to their server nodes. And Rochko himself encouraged users to preemptively block Gab's "instance," resulting in ever fewer users being affected by Gab's attempted invasion of the Mastodon fediverse.
But running a decentralized system creates an entirely new set of moderation issues, which has turned Mastodon itself into a moderation target. Roughly a year after the Gab "invasion," Google threatened to pull Mastodon-based apps from its store for promoting hate speech, after users tried to get around the Play Store ban by creating apps that pointed to Mastodon “instances” filled with hateful content. Google ultimately decided to leave Mastodon-based apps up, but appears ready to pull the trigger on a ban in future.
Originally posted to the Trust & Safety Foundation website.
Thank you for reading this Techdirt post. With so many things competing for everyone’s attention these days, we really appreciate you giving us your time. We work hard every day to put quality content out there for our community.
Techdirt is one of the few remaining truly independent media outlets. We do not have a giant corporation behind us, and we rely heavily on our community to support us, in an age when advertisers are increasingly uninterested in sponsoring small, independent sites — especially a site like ours that is unwilling to pull punches in its reporting and analysis.
While other websites have resorted to paywalls, registration requirements, and increasingly annoying/intrusive advertising, we have always kept Techdirt open and available to anyone. But in order to continue doing so, we need your support. We offer a variety of ways for our readers to support us, from direct donations to special subscriptions and cool merchandise — and every little bit helps. Thank you.
–The Techdirt Team
Filed Under: content moderation, decentralized, federated social media, instances
Companies: gab, mastodon
Reader Comments
Subscribe: RSS
View by: Time | Thread
As a Masto user who was around when the Gab transition went down: Yeah, the reaction from a large number of instances was “fuck all that then” and a near-immediate block on Gab (and Gab-based instances). Gargamel—sorry, Gargron didn’t need to institute a block on Gab; the Fediverse largely did that job on its own.
A similar “job” continues to this day with the #fediblock hashtag, which alerts users across the Fediverse to other users and instances that violate what are considered the general norms of said Fediverse. Since the Fediverse is largely queer-friendly and (at least in spirit, if not in effect) anti-racist, you can imagine what kind of assholes get mentioned in #fediblock posts.
[ link to this | view in chronology ]
Also Diaspora*
Similar issues have arisen with Diaspora* pods, also part of the Fediverse.
Plus there have been individual toxic users who have repeatedly signed up for new accounts on multiple different pods, several per day, in order to get around moderation. The most toxic individual frequently fakes the identity of people he dislikes when signing up at new pods. This makes open registration problematic and raises difficult questions about cross-pod reputation management.
[ link to this | view in chronology ]
An impossibility Lemma
I always thought the Gabs of the world would end up in something like Mastodon -- they want to support generally obnoxious, racist content, they are going to find many businesses and software suppliers refusing to do business with them, so they will have to go to those that allow anonymous, unpaid relationships. That's open source! (And a common carrier ISP result if you start talking about individuals not being able to rent hosting space and therefore working off of their individual internet connections)
As to blocking policy, well, that has to be largely up to the other, independent mastodon instances -- the difficulty arises, of course, when a very considerate friend on a favorite website or mastodon instance says "Look what horrible things they are doing on the Gab mastodon instance! (instructions on how to find it)", or perhaps more indirectly, "The Gab Mastodon instance is doing horrible things...check out Techdirt!(link)"
Context is everything!
[ link to this | view in chronology ]
Wordfilters and the Content Warning system built into Mastodon largely alleviate such issues. Posts using the #fediblock hashtag will call out instances by their domain name, but such posts rarely offer a direct link to those instances, and offending instances are never called out without context.
[ link to this | view in chronology ]
My takeaway:
Gabbers have not yet figured out how to make their own gabiverse list system.
I mean, besides the usual "i need to shove my poop in places it isn't wanted" issue. Never happy just feeding it to them that love it.
[ link to this | view in chronology ]
Re: Gabbers
These GAB persons do not seem to be the most tech wise around when it comes to RTFM. Thank goodness their tech skills are NOT matched by their hatred and fake patriotism which seem to reign supreme. :)
[ link to this | view in chronology ]
If Google could ban Mastodon software for allowing people to connect to instances they don't like, would they also ban web browsers for allowing people to visit websites they don't like?
[ link to this | view in chronology ]
Re:
How might Google accomplish this "banning" of software, Mastodon or browser?
[ link to this | view in chronology ]
Re:
If by that you mean their own browser then sure, they probably could do that, but given it's trivial to swap browsers and there's plenty of competition out there it wouldn't really do much.
They didn't threaten to ban Mastadon software, they threatened to pull the app from their app store.
[ link to this | view in chronology ]
Re:
No problem, it is Android. Just sideload the app and move on.
[ link to this | view in chronology ]
Ultimately the target should be the awful individuals congregating over Gab instance and the ones running the infrastructure behind it. When somebody uses a gun to commit crimes we go after the gun manufacturer or the reseller? Unless they violate some law to sell the weapon to the hipothetical criminal they shouldn't be targeted. Same with mastodon. They cannot be responsible for how people use tools that are completely open source and easily replicated but the host for Gab's instance can be targeted if they knowingly host such speech.
[ link to this | view in chronology ]