Roblox removed over 100 violent games after safety concerns from parents worldwide.
What’s happening: Following the September 10 shooting of conservative commentator Charlie Kirk at Utah Valley University, users on the children’s gaming platform Roblox created violent content including an “assassination simulator” depicting the shooting. The platform, which has 380 million monthly users (nearly 40% under age 13), removed over 100 related games, but the content was initially accessible to children as young as 5.
Why this matters to all parents: This highlights ongoing global challenges with user-generated content platforms where children create and share games. Similar content moderation issues affect platforms worldwide, from Minecraft servers to YouTube Kids, as companies struggle to monitor billions of pieces of user-created content in real-time.
The bigger picture: This represents the fundamental challenge facing all platforms that combine user-generated content with young users. From TikTok’s algorithm concerns to Discord’s chat monitoring, platforms worldwide are grappling with how to balance creative freedom with child safety when users can create and share content instantly.
Here’s what happened with Roblox’s content moderation, how parents worldwide are responding, and what this reveals about children’s gaming platform safety globally.
What Parents Need to Know
What actually happened: Following the September 10 shooting of conservative commentator Charlie Kirk at Utah Valley University, Roblox users created games including an “assassination simulator” that depicted the shooting. The company removed over 100 related games after they were reported, but the content was initially accessible to users of all ages.
What this means for your family: Your child could potentially encounter inappropriate content on Roblox that references real-world violent events before the company’s moderation systems catch it. The platform relies heavily on users reporting problems rather than preventing harmful content from appearing in the first place.
How Roblox’s safety systems actually work: The platform uses automated detection and user reporting to identify problematic content, but this happens after publication rather than before. Children can report inappropriate content using the “Report Abuse” feature, but they may already have been exposed to it.
The age verification reality: Roblox doesn’t require proper age verification beyond entering a date of birth. Games labelled “18 and older” can still be accessed by younger users, and safety features vary based on what ages users claim to be rather than verified ages.
What regulators are doing: Oklahoma’s Attorney General has announced an investigation into Roblox’s safety measures, joining concerns raised by regulators in multiple countries about user-generated content platforms and child safety.
What Regulators Are Saying
Oklahoma Attorney General Gentner Drummond announced an investigation into whether the state can take legal action against Roblox, saying the platform “lacks adequate safety measures and is overrun with harmful content and child predators.”
The investigation follows growing regulatory concern in multiple countries about user-generated content platforms and child safety, with this incident highlighting how quickly inappropriate content can appear before moderation systems catch it.
How This Affects Your Family
If your child uses Roblox: You’ll want to review their recent gaming activity and discuss what to do if they encounter disturbing or inappropriate content. The platform’s safety settings can limit chat functions and restrict certain content, but these must be actively enabled by parents.
If your child uses other platforms: Similar content moderation challenges exist on Minecraft servers, Discord gaming communities, and other platforms where users create and share content. The same vigilance applies across all user-generated content platforms.
Understanding global platform policies: Content moderation standards vary by country, but most major platforms apply similar policies globally. Understanding how these systems work helps parents make informed decisions regardless of location.
Conversation starters with your child:
- “If you see something in a game that makes you uncomfortable, what should you do?”
- “Why do you think some people create inappropriate content on kids’ platforms?”
- “What’s the difference between games made by companies versus games made by other users?”
Warning signs to watch for: Changes in behaviour after gaming, reluctance to discuss what they’ve been playing, or secretive behaviour around gaming activities may indicate exposure to inappropriate content.
Source: The Washington Times: Roblox under scrutiny over disturbing content



