The social networking startup Bluesky, which is working on a decentralized alternative to the platform formerly known as Twitter (now X), gave an intriguing update this Wednesday about its ongoing efforts to ensure user trust and safety. The company is in various stages of developing and testing several initiatives aimed at tackling issues such as harassment, spam, fake accounts, video safety, and more.
To combat malicious actors and harassers, Bluesky is designing new tools to identify when multiple new accounts are created and managed by a single individual. This could significantly reduce instances of harassment where perpetrators use various fake personas to target victims.
In another innovative move, Bluesky is experimenting with detecting and flagging “rude” replies to be reviewed by server moderators. Much like Mastodon, Bluesky is fostering a network where self-hosters and other developers can manage their own servers that interconnect with Bluesky’s main server and others across the network. Though still in early access, this federation feature will eventually empower server moderators to take appropriate action against users who post rude replies. Bluesky plans to diminish the visibility of such replies in its app and will impose account-level sanctions or suspensions on repeat offenders.
To address harassment through lists, Bluesky will remove individual users from a list if they block the list’s creator. This functionality extends to Starter Packs, which are shareable lists that assist new users in finding people to follow on the platform. Bluesky will also monitor for lists with abusive names or descriptions, preventing users from harassing others by adding them to public lists with harmful titles or summaries. Violations of Bluesky’s Community Guidelines will result in the list being hidden until compliant changes are made. Repeat offenders will face further consequences, though detailed plans are still under discussion.
Bluesky is also enhancing its moderation processes, shifting from email-based reports to in-app notifications for handling moderation issues.
In a bid to curb spam and fake accounts, Bluesky is launching a pilot program designed to automatically identify fraudulent or spam accounts. The objective is to respond to these reports within seconds, ensuring swift action against problematic accounts.
One of the standout developments is Bluesky’s strategy to comply with local laws while preserving free speech. The platform will introduce geography-specific labels to hide content in certain regions, keeping in line with local regulations. This approach aims to balance legal compliance with the promotion of free expression.
Additionally, with the recent introduction of video content, Bluesky is implementing features like the ability to turn off autoplay, clear labeling of videos, and video reporting options. The company is keen on user feedback to address any further safety concerns around video content.
Bluesky’s overarching moderation strategy focuses on the frequency and harm of issues. It prioritizes tackling high-frequency and high-damage problems while also keeping an eye on rare but potentially severe edge cases. Even if these issues impact a small number of users, their significant ongoing harm warrants intervention.
Users can voice their concerns through reports, emails, and mentions to the dedicated @safety.bsky.app account.



