This is a follow-up from my previous thread.
The thread discussed the question of why people tend to choose proprietary microblogging platfroms (i.e. Bluesky or Threads) over the free and open source microblogging platform, Mastodon.
The reasons, summarised by @[email protected] are:
- marketing
- not having to pick the instance when registering
- people who have experienced Mastodon’s hermetic culture discouraging others from joining
- algorithms helping discover people and content to follow
- marketing
and I’m saying that as a firm Mastodon user and believer.
Now that we know why people move to proprietary microblogging platforms, we can also produce methods to counter this.
How do we get “normies” to adopt the Fediverse?
@dch82 first, “normies” have to not get harassed when they come here.
Unfortunately the biggest Fedi software refuses to add automated reporting of offensive posts so if it’s not reported, the admins won’t even see it.
People coming from corporate social media are used to ignoring the report button because in their experience, it either doesn’t work, or gets ignored by admins anyway.
We need automated reporting.
@fediverse
Federated reporting would help too
I’m fine with auto REPORTING, but the actual moderation needs to be a human. Auto moderation is bad. It gets things wrong. It’s how I got banned from both twitter (calm down, this was back in 2018 before it was an elon owned nazi cesspool), and reddit.
On twitter I saw a funny video that was posted, and I replied “Aw man, that killed me”.
I was banned for “inciting death threats”
@Lost_My_Mind yeah, just reporting.
I want to do the actual judgement, but if I don’t know the post exists, I can’t judge anything and it makes me so mad that possible racist stuff can exist on my instance without my knowledge because I havent “seen” it.
@fediverse
That’s the thing about automation and training models.
First, they implement some sort of auto-reporting bot that requires a human to review them. In the beginning, it only about 50% accurate, but as they give it more and more examples of good and bad results through the human reviews, it moves to 80%, then 90%, then 99%, then 99.99% accuracy.
After a while, the humans on the other end are so numb to the 9999 entries they have to mark as approved that they can barely tell what’s a rejection themselves, and the moderation team is asking itself just what this human review is actually doing. If it’s 99.99% accurate, why not let the bot decide?
Then, the model moves on from auto-reporting to auto-moderation.
@[email protected] @[email protected] Maybe im a little lost. Isn’t there a block and report button on Mastodon? I’m using Misskey and both buttons seem to work. I mean im reporting to myself, but the button seems to work. What kind of automated blocking are you trying to do here?
@AterNox @dch82 blocking and reporting work fine.
However, people from corporate social media won’t report posts because in their experience, it either doesn’t get taken seriously or the admins ignore it. Corporate social media sites don’t exactly act on reports in a timely manner.
I’m on my own instance, I moderate for myself. I don’t want slurs to exist on my instance at all. However, if I don’t see them with my own eyes, I cannot ban the user.
PS. I’m talking about banning users that are harassing others on the instance level. These are user actions. I am an admin. I run my own instance.
@fediverse
@[email protected] @[email protected] So Mastodon not have a wordlist you can populate that “removes” posts with the keywords you provide? It took me a while to find it in Misskey, works like a charm,
@AterNox @dch82 doesn’t exist for admins. It works on a “user” level. But that won’t remove the post or data from the instance, it just “hides” it so the single user can’t see.
@fediverse
I’m confused, do you mean like automated enforcement rules/algorithms like big SM has? I.e. if user gets reported for breaking Y rule X amount of times ban user for Z amount of time and forward to admin for further action?
@cm0002 no, I want automated reports.
A user using the n word, full on with the hard R, isn’t gonna be a good post. It should be automatically reported to me so that I can judge context and take action.
If a user doesn’t report it, I won’t see it.
I’m on my own instance, I am the user.
If I don’t report it, nobody sees it.
That’s dumb.
@fediverse
Ah, makes sense now, that is dumb. I can totally see why they would have issues with automated enforcement, but what you described I don’t see why anyone would be against it lol
I unironically think it would be easier to train users that the report button works now than it would to get automated reporting that was worth a damn implemented.
By automated reporting do you mean something like filters on the backend to flag offensive posts per some custom settings?