Australia’s internet regulator has accused the world’s largest social media companies of failing to properly enforce the country’s ban on under-16s using their platforms, despite legislation that came into force in December. The eSafety Commissioner, Julie Inman Grant, has raised “serious concerns” about adherence by Facebook, Instagram, Snapchat, TikTok and YouTube, citing poor practices including permitting prohibited users to make repeated attempts at age verification and insufficient measures to stop new account creation. In its initial compliance assessment since the prohibition came into force, the regulator found numerous deficiencies and has now shifted from observation to active enforcement, cautioning that platforms must demonstrate they have implemented “appropriate systems and processes” to prevent children under 16 from accessing their services.
Compliance Failures Uncovered in First Major Review
Australia’s eSafety Commissioner has documented a troubling pattern of failure to comply among the world’s largest social media platforms in her inaugural review since the ban took effect on 10 December. The report shows that Meta, Snap, TikTok, YouTube and Snapchat have jointly failed to implement sufficient safeguards to prevent minors from using their services. Julie Inman Grant raised significant concerns about structural gaps in age verification processes, noting that some platforms have allowed children who initially declared themselves under 16 to later assert they were older, effectively circumventing the law’s intent.
The findings demonstrate a significant escalation in the regulatory response, with the eSafety Commissioner transitioning from monitoring towards direct enforcement. The regulator has emphasised that simply showing some children still hold accounts is inadequate; platforms must instead provide concrete evidence that they have put in place comprehensive systems and procedures designed to prevent under-16s from opening accounts in the outset. This shift signals the government’s commitment to ensure tech giants responsible, with possible sanctions looming for companies that fail to meet the statutory obligations.
- Enabling formerly prohibited users to re-verify their age and restore account access
- Permitting repeated attempts at the identical verification process without penalty
- Inadequate mechanisms to block accounts for under-16s from being created
- Insufficient reporting tools for parents and members of the public
- Absence of clear information about regulatory measures and account removals
The Scope of the Challenge
The substantial scale of social media activity amongst young Australians underscores the compliance challenge facing both the government and the platforms in question. With millions of accounts already removed or restricted since the implementation of the ban, the figures provide evidence of extensive early non-compliance. The eSafety Commissioner’s conclusions suggest that the operational and technical barriers to enforcing age restrictions have proven far more complex than expected, with platforms having difficulty to distinguish genuine age declarations from fraudulent ones. This complexity has placed enforcement authorities grappling with the core issue of whether current age verification technologies are sufficient for the purpose.
Beyond the operational challenges lies a broader concern about the readiness of companies to place compliance ahead of user growth. Social media companies have long resisted strict identity verification requirements, citing data protection worries and the genuine difficulty of confirming age online. However, the regulatory report suggests that some platforms might not be demonstrating adequate commitment to deploy the infrastructure mandated legally. The shift towards active enforcement represents a pivotal moment: either platforms will substantially upgrade their compliance infrastructure, or they stand to incur significant penalties that could transform their operations in Australia and possibly affect compliance frameworks internationally.
What the Figures Indicate
In the first month subsequent to the ban’s implementation, Australian authorities indicated that 4.7 million accounts had been suspended or deleted. Whilst this statistic initially seemed to show compliance achievement, further investigation reveals a more layered picture. The sheer volume of account removals implies that many under-16s had managed to establish accounts in the beginning, indicating that preventive controls were lacking. Additionally, the data prompts inquiry about whether deleted profiles reflect genuine enforcement or just users removing their profiles of their own accord in reaction to the latest limitations.
The restricted transparency surrounding these figures has frustrated independent observers trying to determine the ban’s true effectiveness. Platforms have disclosed minimal information about their enforcement methodologies, performance indicators, or the nature of removed accounts. This opacity makes it difficult for regulators and the wider public to assess whether the ban is functioning as designed or whether younger users are merely discovering alternative ways to use social media. The Commissioner’s push for detailed evidence of systematic compliance measures reflects growing frustration with platforms’ reluctance to provide full information.
Sector Reaction and Pushback
The major tech platforms have responded to the regulator’s enforcement action with a mixture of compliance assurances and doubts regarding the practical feasibility of the ban. Meta, which runs Facebook and Instagram, stressed its dedication to adhering to Australian law whilst simultaneously arguing that accurate age determination remains a major challenge across the industry. The company has called for a different approach, proposing that strong age verification systems and parental consent requirements put in place at the app store level would be more effective than enforcement at the platform level. This stance reflects broader industry concerns that the existing regulatory system puts an impractical burden on individual platforms.
Snap, the developer of Snapchat, has taken a more proactive public stance, announcing that it had locked 450,000 accounts following the ban’s implementation and claiming to continue locking more daily. However, industry observers question whether such figures reflect authentic adherence or simply represent reactive account management. The fundamental tension between platforms’ business models—which historically relied on maximising user engagement and growth—and the statutory obligation to systematically remove an whole age group remains unresolved. Companies have consistently opposed rigorous age verification methods, pointing to privacy issues and technical constraints, establishing an impasse between regulators and platforms over who carries responsibility for implementation.
- Meta maintains age verification should occur at app store level instead of on individual platforms
- Snap claims to have locked 450,000 accounts following the ban’s implementation in December
- Industry groups cite privacy concerns and technical challenges as impediments to effective age verification
- Platforms contend they are doing their best whilst challenging the ban’s overall effectiveness
More Extensive Inquiries Concerning the Ban’s Efficacy
As Australia’s under-16 social media ban moves into its implementation stage, fundamental questions remain about whether the law will achieve its intended goals or merely push young users towards unregulated platforms. The regulatory authority’s initial compliance assessment reveals that following implementation, significant loopholes exist—children keep discovering ways to bypass age verification mechanisms, and platforms have had difficulty prevent new underage accounts from being created. Critics argue that the ban’s success depends not merely on regulatory vigilance but on whether young people will truly leave mainstream platforms or simply shift towards other platforms, secure messaging apps, or VPNs designed to conceal their age and location.
The ban’s worldwide effects contribute further complexity to assessments of its success. Countries including the United Kingdom, Canada, and several European nations are watching Australia’s experiment closely, exploring similar regulatory measures for their own citizens. If the ban does not successfully reduce children’s digital engagement or does not protect them from damaging material, it could weaken the case for similar measures elsewhere. Conversely, if regulation becomes sufficiently robust to genuinely restrict underage access, it may inspire other governments to implement similar strategies. The result will probably shape worldwide regulatory patterns for many years ahead, making Australia’s enforcement efforts analysed far beyond its borders.
Those Who Profit and Who Loses
Mental health advocates and organisations focused on child safety have championed the ban as a necessary intervention against algorithmic manipulation and contact with harmful content. Parents and educators contend that taking young Australians off platforms designed to maximise engagement could lower anxiety levels, enhance sleep quality, and reduce exposure to cyberbullying. Tech companies’ own research has recognised the risks to mental health linked to social media use amongst adolescents, adding weight to these concerns. However, the ban also eliminates valid applications of social media for young people—keeping friendships alive, obtaining educational material, and engaging with online communities around common interests. The regulatory approach assumes harm exceeds benefit, a calculation that some young people and their families question.
The ban’s real-world effects extends beyond individual users to influence content creators, small businesses, and community organisations dependent on social media platforms. Young people who might have taken up creative careers through platforms like TikTok or Instagram now encounter legal barriers to participation. Small Australian businesses that are dependent on social media marketing no longer reach younger demographic audiences. Community groups, charities, and educational organisations struggle to reach young people through channels they previously employed effectively. Meanwhile, the ban unexpectedly advantages large technology companies with resources to create age verification infrastructure, possibly reinforcing their market dominance rather than reducing it. These unforeseen effects suggest the ban’s effects extend far beyond the simple goal of child protection.
What Lies Ahead for Regulatory Action
Australia’s eSafety Commissioner has indicated a notable transition from passive monitoring to direct intervention, marking a critical turning point in the implementation of the age restriction. The watchdog will now gather evidence to determine whether platforms have omitted “reasonable steps” to restrict child participation, a legal standard that extends beyond simply documenting that minors continue using these platforms. This strategy requires demonstrable proof that platforms have implemented proper safeguards and processes intended to prevent minors. The enforcement team has indicated it will launch probes systematically, developing arguments that could result in substantial penalties for non-compliance. This transition from observation to intervention reveals mounting concern with the platforms’ current efforts and signals that voluntary cooperation on its own will not be enough.
The implementation stage highlights critical issues about the sufficiency of sanctions and the operational systems for maintaining corporate responsibility. Australia’s statutory provisions delivers regulatory tools, but their efficacy depends on the eSafety Commissioner’s willingness to pursue formal action and the platforms’ capability to adjust meaningfully. Overseas authorities, especially regulators in the Britain and Europe, will carefully track Australia’s regulatory approach and results. A successful enforcement campaign could establish a template for further jurisdictions contemplating equivalent prohibitions, whilst failure might weaken the comprehensive regulatory system. The coming months will prove crucial whether Australia’s pioneering regulatory approach produces genuine protection for teenagers or becomes largely performative in its impact.
