Aaron Rodericks
@aaron.bsky.team
25K followers 1.1K following 920 posts
Canadian wanderer in Ireland. Trying to make the internet a better place. Bluesky Head of Trust and Safety. Email [email protected] for more complex issues. Priority notifications active, so I don't see mentions. abuse = block
Posts Media Videos Starter Packs
aaron.bsky.team
Tbh I believe most of your concerns are with content moderation as a system. Some of this can be improved as we mature as a tiny startup, but fundamentally it would take a longer explanation than I can cover in posts to cover why it’s so difficult to make decisions at scale. I’ll send a ping.
Reposted by Aaron Rodericks
bsky.app
Bluesky @bsky.app · Nov 15
Bluesky uses AI internally to assist in content moderation, which helps us triage posts and shield human moderators from harmful content. We also use AI in the Discover algorithmic feed to serve you posts that we think you’d like.

None of these are Gen AI systems trained on user content.
aaron.bsky.team
Every appeal gets reviewed by moderators, but literally no social media site at scale can scan user content in a timely fashion for violations using only humans. They do not retain or train on any Bsky user data or images.
aaron.bsky.team
It's an automated system, so unfortunately this is an error that gets made surprisingly often with prosthetics.
aaron.bsky.team
the most recent one was a temp error on asukafield where a mod hid the list and it was rapidly reversed on invesigation. Is that the one you are referencing, or just lists in general?
aaron.bsky.team
It was a brand new bot that we had just activated to deal with the massive backlog in reports after the US election.

We currently have a much smaller backlog and mods are behind on handling reports, since we had a significant surge in reports that we are still catching up on.
aaron.bsky.team
TBH that was an error on our end that I wasn't aware of until extremely recently pfrazee.leaflet.pub/3lz4sgu7iec2k. It's being addressed architecturally. As far as I'm aware we only ban CSAM at a relay level, but I'm not particularly technical.
Update on Protocol Moderation - Paul's Leaflets
Where account takedowns happen is important
pfrazee.leaflet.pub
aaron.bsky.team
yes we had an impersonation bot that assumed the account was fake upon signup (because people had created similar fake account prior so it focused on that pattern). That was totally unrelated to any assessment under the community guidelines.
aaron.bsky.team
That wording is in our community guidelines. bsky.social/about/suppor... It's still on our site and they expire on Oct 15 2025. So seeing as the wording is still live and on the site, how were they changed last December?
Community Guidelines (Deprecated) - Bluesky
bsky.social
aaron.bsky.team
cause people raise it in my mentions daily and literally won't stop, so I might as well share my side.
aaron.bsky.team
We know it’s frustrating to see bad behaviour elsewhere. But trying to police the entire internet isn’t possible for any team — and making exceptions about when to do so only adds confusion and inconsistency. 9/9
aaron.bsky.team
This approach keeps our system scalable and fair as we grow. Every decision must be explainable, repeatable, and based on evidence our tools can reliably evaluate. We’re also working to make our decisions clearer and more transparent. 8/9
aaron.bsky.team
At Bluesky, moderation focuses on behaviour we can observe here: harassment, threats, hate, spam, impersonation, and other violations of our Community Guidelines. 7/9
aaron.bsky.team
When moderation expands into other apps or private servers, we risk huge workloads, privacy issues, and inconsistent enforcement. That’s why nearly every platform limits enforcement to what happens on their own service. 6/9
aaron.bsky.team
Screenshots can be edited, context can be missing, and identities are hard to confirm across platforms. Moderation depends on repeatable, auditable processes — not assumptions based on unverifiable off-platform material. 5/9
aaron.bsky.team
Let’s say someone sends us screenshots or Discord logs — 15,000 lines of conversation — claiming they prove coordinated harassment (this happened). That’s hours or days of work to verify and still might not show any clear link to behaviour on Bluesky. 4/9
aaron.bsky.team
Right now, we’re a small team with a growing backlog. The slower we get, the more people wonder why decisions aren’t being made in real time — but with human review and millions of posts, that’s simply not possible. 3/9
aaron.bsky.team
A single moderator reviews thousands of reports per year — often making a decision in under 30 seconds. Expanding the scope of what counts as “evidence” makes it much harder to act fairly and consistently. 2/9
aaron.bsky.team
People often ask why we don’t act on off-platform evidence when investigating reports. Here’s an example of how complex that can get. 1/9
aaron.bsky.team
That was always the case? I'll post a longer thread explaining why.
aaron.bsky.team
that second was a mod error and she has been released with an apology already. Mods make errors continually with hundred mods reviewing tickets in 30 seconds or less.
aaron.bsky.team
If you want to have a conversation about it, or other mod decisions that you disagree with, sure - I can understand that, but endlessly coming up with a screenshot as evidence of...? isn't particularly constructive.