Meta is implementing new artificial intelligence tools in Australia to detect users under 18 on its platforms (Instagram, Facebook), even if they’ve claimed to be over 18, and automatically place them into “Teen Account” settings with tighter content/interaction restrictions. This move comes as Australia prepares to enforce a world-first law from December 10, 2025 that bans social media accounts for children under 16 under the Online Safety Amendment (Social Media Minimum Age) Act 2024. Meta’s “Teen Account” protections include restrictions on who can contact the user, limits on live streaming, blurred or limited exposure to “sensitive content”, and other age-appropriate guarding features. The decision reflects both legal pressures on social media platforms to comply with upcoming regulation and ethical concerns around youth exposure to harmful content; however, critics warn about privacy risks, the technical challenges of accurately inferring age, and unintended consequences.
Sources: Epoch Times, SmartCompany, Meta
Key Takeaways
– Regulatory pressure is accelerating platform change. Australia has passed a law (Online Safety Amendment 2024) that from December 2025 will ban under-16s from having social media accounts. This forces platforms like Meta to adopt technologies (AI detection, age verification) to comply.
– AI age-detection is being used preemptively, but comes with limitations. Meta is now using AI to flag users who list an adult age but exhibit behavior or patterns more consistent with younger users, shifting them into safer “Teen Account” settings. But distinguishing real age via algorithms is imperfect; false positives/negatives and privacy trade-offs are important.
– Privacy, rights, and implementation are contested. While many see value in protecting minors from harmful content, critics argue the law may overreach, risk privacy, or drive younger users underground or into unregulated spaces. There are also concerns about how “reasonable steps” will be defined, how platforms will enforce the rules, and whether age verification will infringe on civil liberties.
In-Depth
Australia is now moving from policy announcement into active enforcement. The government’s Online Safety Amendment (Social Media Minimum Age) Act 2024, passed in late 2024, changes the legal landscape by making it a requirement for social platforms to prevent users under 16 from maintaining or creating accounts. That law becomes fully enforceable as of December 10, 2025. In response, Meta is rolling out AI-based age detection to identify users who may be under 18 (even if their stated age is older) and rewiring their experiences on Instagram and Facebook into more restrictive “Teen Account” settings. These settings include limitations on who can contact them, restrictions on live streaming, and less exposure to sensitive content. The change aims to reduce exposure to risks typically associated with adolescent use of social media: harassment, exposure to inappropriate content, and mental health stressors.
Meta’s blog post about expanding its Teen Account protections reveals that the company is building in parental engagement—informing parents about why accurate age information matters, how to confirm age, and giving users a way to appeal if they believe the algorithm mis-classified them. Meanwhile, public debate continues over how far regulation should go. Some advocates see the AI approach as necessary, others warn that technological detection of age could be error prone or misused. Privacy groups are concerned about what data is used for age detection, how transparent the criteria are, and what recourse users have if incorrectly flagged.
Another contested issue is the definition of what steps are “reasonable” under the law. The law imposes significant fines for non-compliance, which pushes platforms to adopt stronger age verification tools. But “reasonable” could vary, especially across services that differ in functionality (messaging vs broadcast vs content sharing, etc.). Additionally, loopholes may emerge: users may circumvent bans via VPNs, fake IDs, or undeclared accounts; content may still be accessible via unlogged-in browsing; and enforcement against non-complying platforms or foreign services may prove difficult.
In sum, Australia is setting something of a global precedent in forcing tech platforms to internalize age as a core dimension of user safety. Meta’s AI tools appear designed to bridge the gap between legal mandate and user experience. But achieving that safely—without compromising privacy, over-banning, or creating new harms—will require precision, transparency, and ongoing oversight.

