Meta has been ordered to pay £279m (approximately $375m) by a New Mexico court after a jury determined that the social media company liable for misleading the public about child safety on its platforms. The significant decision marks the first time a state has brought legal action against Meta—which owns Facebook, Instagram and WhatsApp—over allegations that its services endangered children and subjected them to sexually explicit material and contact with sexual predators. New Mexico’s Attorney General Raul Torrez described the verdict as “historic”. Meta, led by chairman and chief executive Mark Zuckerberg, has stated that it disputes with the decision and intends to challenge it, arguing that it works hard to ensure users safe online.
The New Mexico State Ruling and Its Importance
The New Mexico jury’s ruling to find Meta liable for breaching the state’s consumer protection laws represents a significant milestone in the persistent fight over platform responsibility. During a gruelling seven-week trial, jurors were exposed to damaging internal Meta records and evidence from previous workers who revealed the company’s knowledge of predatory individuals abusing its platforms. The magnitude of the breaches—which the jury established reached the thousands—highlights the endemic scope of the difficulties impacting Meta’s platforms. Each violation incurred a highest sanction of $5,000, ultimately totalling the $375m judgment.
The case attracted particular attention after testimony from Arturo Béjar, a former engineering leader at Meta who turned whistleblower after leaving the company in 2021. Béjar detailed experiments he carried out on Instagram showing that underage users received sexualised content, and he related a profoundly intimate account: his own young daughter was propositioned for sex by a stranger on the platform. State prosecutors also presented internal Meta research showing that 16 per cent of all Instagram users had reported encountering unwanted nudity or sexual activity within a single week—a striking statistic that illustrated the pervasiveness of the problem.
- Meta operates Facebook, Instagram, WhatsApp and other major platforms
- Jury identified thousands of violations of New Mexico’s consumer protection laws
- Former employee provided evidence on sexualised content exposed to minors
- Company intends to appeal the significant verdict
How the Court Determined Meta Liable
Internal Documents and Whistleblower Testimony
The prosecution’s case depended significantly on damaging evidence that came from Meta’s own internal operations. Throughout the 7-week trial, jurors reviewed confidential company documents that showed Meta’s knowledge of the risks children encountered on its platforms. These materials proved instrumental in establishing that the company recognised the risks yet failed to adequately protect young users. The evidence painted a picture of a corporation cognisant of systemic problems but unwilling to prioritise child safety over user engagement figures and platform growth.
Central to the prosecution’s case was testimony from Arturo Béjar, whose internal vantage point carried substantial impact with the jury. As a previous head of engineering, Béjar held thorough grasp of how Meta’s systems worked and where safety protocols fell short. His readiness to go on record about his experiences, including the distressing narrative of his own daughter being propositioned for sex on Instagram, provided authenticity and emotional resonance to the state’s claims. His testimony bridged the gap between abstract corporate wrongdoing and real injury to actual minors.
The Magnitude of the Challenge
State prosecutors presented Meta’s own research to illustrate the staggering prevalence of damaging material on its platforms. Internal studies revealed that 16 per cent of all Instagram users had reported experiencing non-consensual sexual content within a single week—a figure that astonished the jury and underscored the normalisation of exploitation across the social media giant’s services. This statistic formed the foundation of the prosecution’s case, illustrating that the problem was not isolated incidents but rather a widespread, systemic failure.
The jury’s conclusion that Meta had committed thousands of violations of New Mexico’s Unfair Practices Act emphasised the prevalence of the matters in question. With each violation incurring a maximum penalty of £5,000, the cumulative total reached £279m. This methodology reflected not merely a single lapse in conduct but rather ongoing, deliberate failures across Meta’s business activities. The considerable quantity of violations demonstrated that child endangerment had become embedded within the company’s operational model rather than constituting sporadic failures.
Meta’s Defence and Ongoing Efforts
Meta has strongly disputed the New Mexico jury’s conclusions, with the company’s spokeswoman emphasising that it “works hard to protect users on our platforms” and remains “confident in our track record on safeguarding teens online.” The platform operator has signalled its intention to appeal the verdict, suggesting it believes the court’s decision was flawed or disproportionate. Meta’s defence throughout the trial centred on the argument that identifying and removing malicious users and dangerous material presents real, fundamental difficulties for platforms functioning across global markets. The company argued that it has made substantial investments in safety features and that the problem of child exploitation, whilst serious, cannot be entirely eliminated through technology by itself.
In recent times, Meta has launched several initiatives intended to managing child safety concerns and possibly limiting harm to its public image. Instagram rolled out Teen Accounts in 2024, giving younger users greater control over their digital activities and restricting access to potentially harmful content. Most recently, the platform introduced a feature intended to inform parents when their children look for self-harm content, constituting an attempt to reconcile the conflict between adolescent privacy and parental control. These steps, though, came in the wake of years of scrutiny and legal action, prompting debate about whether they amount to genuine commitment to safety or defensive public relations in response to prolonged pressure from regulators and the public.
- Instagram Teen Accounts provide improved privacy safeguards for teenage users
- Recently introduced parental alert feature alerts parents to searches related to self-harm
- Meta contends structural difficulties make complete content removal unfeasible
Wider Legal Framework and Industry Implications
The New Mexico verdict represents a turning point in the intensifying conflict between digital authorities and social media giants over safeguarding minors. This is the first occasion on which a state has prevailed against Meta in court proceedings on endangerment charges, setting a landmark ruling that could embolden other authorities to pursue like cases. The $375m penalty, whilst substantial, is dwarfed by Meta’s yearly earnings, yet the deeper meaning cannot be overstated. The case shows that juries are more prepared to enforce corporate accountability for the impacts of their algorithmic recommendation systems and operating practices, especially where internal evidence suggests company awareness of wrongdoing.
Beyond Meta, the implications reverberate across the tech industry. Google, which owns YouTube, encounters comparable allegations in separate litigation, whilst TikTok and other networks confront mounting scrutiny from regulatory bodies and legislators globally. The New Mexico case illustrates how state-based action can circumvent federal regulatory impasse, with enforcement officials utilising consumer protection statutes initially intended for traditional commerce. This patchwork approach may be more successful than expecting comprehensive federal legislation, yet generates ambiguity for tech firms conducting business across various regions with inconsistent regulatory frameworks and compliance focus relating to child safety obligations.
| Jurisdiction | Status |
|---|---|
| New Mexico | Jury verdict: Meta liable, $375m penalty awarded |
| Los Angeles | Separate trial ongoing regarding addiction claims |
| Federal courts | Thousands of similar lawsuits in progress |
| Global regulators | Increasing scrutiny of platform safety measures |
The convergence of state litigation, federal regulatory scrutiny, and international oversight suggests that technology companies encounter an unparalleled reckoning over child protection measures. Whether this New Mexico verdict sparks genuine industry-wide reform or merely represents a passing defeat for Meta is uncertain, but the court decision makes clear that judges are no longer accepting corporate guarantees about safety efforts when company records conflicts with public assertions.
