From 10 December, children under 16 will no longer be able to hold accounts on most major social media platforms. The goal is simple: reduce online risks such as bullying, predators, harmful content, and the rising mental-health impact of excessive screen time.
Which Platforms Are Affected?
Included: Facebook, Instagram, TikTok, Snapchat, YouTube, X (Twitter), Threads, Twitch, Kick, Reddit.
Not included: Messenger Kids, WhatsApp, YouTube Kids, Kids Helpline, Google Classroom.
How Will It Be Enforced?
Platforms—not parents—must verify age.
Possible methods include ID checks, facial or voice recognition, or behaviour-based age estimation. Self-declared ages will no longer count.
Meta has begun locking under-16s out, and Snapchat already supports verification tools. Other platforms will follow.
Key Concerns
- Privacy: Sensitive ID or biometric data may be required.
- Mental health: Sudden loss of online connection may affect teens socially.
- Workarounds: Kids may try to bypass systems.
- Gaps: Gaming platforms like Roblox and Discord aren’t included.
Why It May Help
The restrictions aim to create a safer, healthier digital environment with:
- Reduced exposure to harmful content
- Less cyberbullying
- Improved sleep and wellbeing
- Lower tech dependency
How Parents Can Support Their Child
- Reduce screen time gradually before the cutoff.
- Acknowledge emotions—losing platforms can feel big.
- Check in regularly for signs of withdrawal or stress.
- Offer positive alternatives like outdoor play, hobbies and family time.
- Model healthy habits by showing balanced device use yourself.