Roblox now employs AI moderation to prevent harmful content from reaching you.
Roblox is enhancing its safety measures with a new AI moderation system
If you have ever used Roblox, you understand how chaotic and unpredictable the platform can be. Now, Roblox is leveraging that unpredictability to transform its content moderation approach.
The platform has introduced a real-time multimodal AI moderation system that not only analyzes individual items but also scans complete in-game scenes in real-time to identify content that may have evaded previous checks.
How does Roblox’s new AI moderation function?
In contrast to older moderation tools that assess one element at a time, the new multimodal system evaluates an entire scene, including avatars, text, and 3D objects, to ascertain whether the overall combination violates Roblox’s Community Standards.
For instance, if someone uses a free-drawing tool to create an offensive symbol — which wouldn't be flagged during a single-object review but is clearly problematic in context.
When the system detects repeated violations in a single game instance, it disables only that particular server instead of the entire game. Since its launch, approximately 5,000 servers have been taken down daily.
Roblox aims to monitor 100% of playtime with this system. In addition to server shutdowns, the platform is also creating tools to identify and eliminate individual offenders without negatively impacting others’ experiences.
What does this entail for Roblox creators?
Creators are kept informed as well. A new chart in the current Creator Dashboard displays the number of game servers that were shut down on any given day due to user misconduct.
Increases in this number can indicate a problem that needs attention, allowing creators the opportunity to review and modify aspects like custom emotes or in-game building tools before issues escalate.
Introducing a new certification program for improved online gaming
Roblox is also addressing a broader industry challenge. In collaboration with Keyword Studios and Riot Games, it is co-developing the DLC Leadership Program aimed at online community managers and moderators.
Research psychologist Rachel Kowert, who is the Research Director at Games for Change, is overseeing the academic component. The objective is to create standardized, evidence-based training for individuals managing gaming communities, something that has previously been lacking in the industry.
Roblox has been strengthening its safety measures for some time, from implementing parental controls to age-based chat filters, and this recent update represents its most significant initiative to date.
Other articles
Roblox now employs AI moderation to prevent harmful content from reaching you.
Roblox's updated AI moderation system analyzes entire game scenes in real-time, detecting harmful content that many older systems would overlook, and is already closing down 5,000 servers each day.
