━━━━━━━━━━━━━━━━━━━━━━━━ ⚠️ BREAKING CHILD SAFETY NEWS
This is one of three major child safety developments this week:
- Italian Families Sue Meta & TikTok
- Kentucky Sues Roblox Over Child Predators
- Australia Issues Urgent Warning: Kids Exposed to Graphic Violence ← You’re here ━━━━━━━━━━━━━━━━━━━━━━━━
One in five children has seen graphic real-world violence online – murders, assassinations, mass casualty events – and they’re not looking for it. It’s finding them.
Australia’s eSafety Commissioner issued an urgent Online Safety Advisory on October 6, 2025, after research revealed that 22% of children between ages 10 and 17 have been exposed to extreme violence online. Not movie violence. Not video game violence. Real people dying in real-world events, algorithmically served to children’s devices.
This isn’t about teenagers deliberately seeking out shocking content. It’s about platforms pushing graphic reality to children through autoplay, recommendations, and shares – often before they can look away.
What Australia’s Research Found
The eSafety Commissioner’s research painted a disturbing picture of what children are encountering online:
22% of kids aged 10-17 have seen extreme real-life violence, including:
- Recent political assassinations
- Brutal murders caught on camera
- Mass casualty events and terrorist attacks
- Graphic conflict and war footage
- Execution videos
- Severe injuries and deaths
This content isn’t hidden in dark corners of the internet. It’s appearing on mainstream platforms that children use every day: X (formerly Twitter), Facebook, Instagram, Snapchat, TikTok, and YouTube.
So-called “gore” content is reaching children through:
- Autoplay features: Videos start playing before children realize what they’re watching
- Algorithm recommendations: “You might also like” suggestions based on previous viewing
- Direct messages: Sent by peers or strangers
- Reposts and shares: Content spreads through networks, reaching unintended audiences
The Commissioner’s urgent advisory emphasized that this isn’t isolated incidents – it’s a systemic problem requiring immediate parental action and platform accountability.
How This Content Reaches Children
Understanding how graphic violence finds its way to children helps parents protect against it:
Algorithmic Amplification
Platforms use algorithms to keep users engaged. If a child watches any intense or dramatic content – even accidentally – the algorithm may interpret this as interest and serve similar (or more extreme) content.
How it escalates:
- Child sees news clip about a dramatic event
- Algorithm notes the engagement (even a few seconds of viewing)
- More intense news content appears in recommendations
- Graphic “raw footage” versions start appearing
- Extreme violence enters the feed
The algorithm doesn’t distinguish between “watched because interested” and “watched because shocked.” It just sees engagement.
Autoplay Features
Many platforms automatically play the next video without user action. A child watching innocent content can suddenly encounter graphic violence when the next video autoplays.
Why autoplay is dangerous:
- No warning or content description before it starts
- Often happens when device is left playing unattended
- Difficult to stop quickly enough to avoid exposure
- Creates unexpected trauma
Social Sharing
When graphic content goes viral, it spreads through social networks rapidly. Well-meaning friends might share shocking content with “OMG did you see this??” messages, not realizing the psychological impact.
The share cycle:
- Someone posts graphic real-world violence
- It’s shared as “important to see” or “can you believe this?”
- Each share reaches new networks, including children
- Platforms’ content moderation can’t keep up with viral spread
- By the time it’s removed, millions have seen it
Direct Messages
Predators or bullies sometimes send graphic content directly to children as a form of harassment or to desensitize them to disturbing material.
Which Platforms Are Involved
Australia’s advisory specifically named platforms where this content is appearing:
X (Twitter):
- Graphic content often shared as “breaking news”
- Community Notes system can’t moderate fast enough
- Algorithm promotes high-engagement (shocking) content
Facebook:
- Age verification easily bypassed
- Content shared in groups and through messenger
- Autoplay in feeds exposes users to unwatched content
Instagram:
- Stories and Reels can contain graphic footage
- Explore page algorithm recommends based on engagement
- Direct messages used to share shocking content
Snapchat:
- Discover section shows news content, some graphic
- Direct snaps can contain disturbing material
- Disappearing messages make reporting difficult
TikTok:
- For You page algorithm serves content based on engagement
- Live streams can show real-time disturbing events
- Duets and stitches spread content rapidly
YouTube:
- Autoplay feature moves from innocent to disturbing content
- News channels post graphic footage
- Algorithm recommends increasingly intense content
Why This Isn’t Just an Australia Problem
The research comes from Australia, but the platforms involved operate globally with the same algorithms, the same autoplay features, the same content moderation challenges.
There’s no reason to believe children in other countries aren’t seeing similar content.
In fact, Australia’s eSafety Commissioner is considered a world leader in online safety research. Other countries often adopt Australian findings as indicators of global trends.
What this means for parents worldwide:
- The 22% statistic likely applies to your country too
- The same platforms are serving the same algorithmic feeds globally
- Your child’s device probably has autoplay enabled by default
- Current content moderation isn’t working anywhere
The Psychological Impact on Children
Exposure to graphic real-world violence affects children differently than fictional violence:
Why real violence is more traumatic:
- Children know it actually happened to real people
- They may identify with victims or fear similar events
- Lack of narrative context (unlike movies with beginnings, middles, ends)
- Often unexpected, creating shock and helplessness
Potential psychological effects:
- Intrusive thoughts and images
- Nightmares and sleep disturbances
- Increased anxiety about safety
- Desensitization to violence
- Difficulty distinguishing reality from online content
- Fear of similar events happening to them or loved ones
Long-term concerns: Research on children exposed to graphic content shows potential for:
- PTSD-like symptoms
- Difficulty regulating emotions
- Changed perception of the world as dangerous
- Normalized violence as acceptable
What Parents Should Do Now
Australia’s urgent advisory came with implicit recommendations for immediate parental action:
Check Devices Yourself
Don’t just ask if they’ve seen anything disturbing – look.
Many children don’t report graphic content because they:
- Feel shocked and don’t know how to process it
- Fear losing device access if they admit what they saw
- Think it’s “normal” because the algorithm keeps showing it
- Feel embarrassed or responsible
How to check:
- Scroll through their social media feeds with them present
- Look at recommended content and suggestions
- Check direct messages and group chats
- Review watch/view history where available
Disable Autoplay Features
Stop content from playing without permission:
YouTube:
- Settings → Autoplay → Turn off
TikTok:
- Settings → Accessibility → Turn off autoplay
Instagram:
- Settings → Account → Cellular data use → Use less data (disables autoplay)
Facebook:
- Settings → Media → Never autoplay videos
X (Twitter):
- Settings → Accessibility → Video autoplay → Never
Have Honest Conversations
Talk to your children about what to do when disturbing content appears:
“If you see something violent, scary, or disturbing online:
- Look away immediately – don’t keep watching
- Don’t share it – even to warn friends
- Tell a trusted adult – parent, teacher, counselor
- Know it’s not your fault – algorithms push this content”
Age-appropriate conversation starters:
For younger children (10-12): “Sometimes scary or upsetting things from the news can show up on apps. If you see something that makes you feel scared or worried, close the app and come tell me right away. You won’t be in trouble.”
For teens (13-17): “I know you’ve probably seen disturbing content online – a lot of kids have. The way algorithms work, they can show you graphic real-world violence even if you’re not looking for it. If this happens, you can always talk to me about it. We can process it together.”
Review Platform Settings Together
Make safety changes as a family activity, not punishment:
- Enable Restricted Mode / Sensitive Content filters
- Adjust privacy settings to limit who can contact them
- Turn off location services
- Review and limit app permissions
- Set screen time boundaries
Important: Explain WHY you’re making changes. When children understand that algorithms push disturbing content, they’re more likely to accept safeguards.
Know the Warning Signs
Watch for signs your child has been exposed to disturbing content:
Behavioural changes:
- Sudden reluctance to use devices
- Nightmares or sleep problems
- Increased anxiety or fearfulness
- Avoiding news or current events discussions
- Asking unusual questions about violence or death
Emotional responses:
- Seems upset after device use
- Withdraws from family
- Shows signs of distress when certain topics arise
- Exhibits anger or aggression
If you notice these signs:
- Create a safe, non-judgmental space to talk
- Ask open-ended questions: “You seem worried lately. Want to talk about what’s on your mind?”
- Consider professional support if distress continues
What Needs to Change
Australia’s urgent advisory implicitly calls for systemic changes:
Platforms must:
- Disable autoplay by default for users under 18
- Implement stronger content warnings before graphic material
- Improve algorithm design to stop recommending violent content to children
- Speed up content moderation for graphic real-world violence
- Make parental controls easier to find and use
Governments should:
- Require platforms to prove safety measures work
- Mandate age-appropriate design for services used by children
- Hold platforms accountable for algorithmic harm
- Invest in education about online safety
Parents need:
- Better tools to monitor and control content
- Clear information about platform risks
- Support in helping children process disturbing content
- Resources for when children are affected
The Broader Context
This advisory comes amid growing global concern about children’s online safety:
This same week:
- Italian families sued Meta and TikTok over child safety failures
- Kentucky sued Roblox over predators and abuse material
- Multiple countries are implementing age restrictions for social media
The pattern is clear: Platforms designed for adult engagement are serving adult-level disturbing content to children, and current safeguards aren’t working.
What You Can Do Beyond Your Family
Report problems:
- Report graphic content through platform reporting features
- Document patterns of inappropriate recommendations
- Share concerns with school administrators
Advocate for change:
- Contact your government representatives about platform accountability
- Support organizations working on child online safety
- Join parent advocacy groups
Educate your community:
- Share this information with other parents
- Discuss at parent-teacher meetings
- Help other families understand the risks
The Bottom Line
Your child doesn’t have to seek out disturbing content to find it. Algorithms, autoplay, and viral sharing push graphic real-world violence to children’s devices every day.
22% of children have already seen extreme violence online. That’s more than one in five kids – possibly including yours.
This isn’t about banning devices or eliminating all online access. It’s about understanding that platforms built for adult engagement are fundamentally unsuited for children without significant safeguards.
Australia issued an urgent advisory for a reason: this is happening now, it’s affecting millions of children, and parents need to act.
Check your child’s device today. Have the conversation tonight. Make the setting changes this week.
Because the algorithms won’t wait, and neither should you.
Related Reading
- Italian Families Sue Meta & TikTok Over Child Safety Failures
- Kentucky Sues Roblox Over Child Predators and Sexual Abuse Material
Has your child been exposed to graphic content online? How did you handle it? Share your experience to help other parents.
Want to stay informed about child safety developments?
Get Plugged In – what every parent in today’s digital world needs to know. Delivered free every Thursday.
www.wired-parents.com/subscribe
Sources: Medianet, eSafetyCommissioner



