Source of this article and featured image is TechCrunch. Description and key fact are generated by Codevision AI system.

Bluesky, a decentralized social network, has introduced new moderation tools to enhance transparency and community safety. The updates include expanded reporting categories, a severity-based strike system, and improved enforcement tracking. Legal pressures, such as the UK’s Online Safety Act, drive these changes to protect minors online. A recent incident saw user Sarah Kendzior suspended for a comment interpreted as a threat, highlighting the platform’s strict enforcement. Bluesky aims to balance community freedom with accountability while navigating growing regulatory demands.

Key facts

  • Bluesky expanded reporting options from six to nine categories to address specific issues like youth harassment and human trafficking.
  • The strike system now assigns severity ratings to violations, determining penalties from temporary suspensions to permanent bans.
  • Legal requirements, including the UK’s Online Safety Act, influence Bluesky’s moderation policies to protect minors and comply with regulations.
  • User Sarah Kendzior was suspended for a comment deemed a threat, reflecting the platform’s strict enforcement of community guidelines.
  • Bluesky faces criticism for not banning controversial users, despite its focus on balancing free speech with platform safety.
See article on TechCrunch