Meta faces fines and legal battles over child safety, social media addiction, and AI delays—Facebook’s parent company is under heavy scrutiny.
As of April 2026, Turkey has passed a law banning children under 15 from creating social media accounts, requiring platforms to install age-verification systems. The UK government is debating a similar ban for under-16s amid parliamentary disagreements. Australia has enforced a ban since December 2025, with other countries like Spain, France, and Indonesia considering or implementing similar restrictions.
On February 9, 2026, New Mexico begins the first stand-alone trial against Meta over alleged harms to children on its platforms. The case involves undercover investigations, claims of algorithm-driven addiction, and failure to disclose harmful effects, with Meta denying violations. The trial could influence future legal actions nationwide.
As of February 2026, a landmark trial in Los Angeles County Superior Court is underway against Meta and Google's YouTube, accused of deliberately designing platforms to addict children and harm their mental health. The case centers on a 19-year-old plaintiff, KGM, and could set precedent for hundreds of similar lawsuits. TikTok and Snap settled earlier. Meta denies wrongdoing, citing complex mental health factors and safeguards.
A New York neighborhood faces tension after a resident is accused of deliberately covering a neighbor's car with snow during a blizzard. The incident, captured on video, has gone viral, highlighting ongoing neighbor disputes and issues of harassment amid severe winter conditions.
As of March 6, 2026, Indonesia has enacted a regulation banning children under 16 from having accounts on high-risk social media platforms including YouTube, TikTok, Facebook, Instagram, Threads, X, Roblox, and Bigo Live. The ban will be implemented gradually starting March 28, with penalties for non-compliant platforms. This makes Indonesia the first Southeast Asian country to impose such restrictions, following Australia's December 2025 ban.
Meta, the owner of Facebook and Instagram, is being sued by New Mexico prosecutors for allegedly failing to disclose known risks of social media addiction and child exploitation. The trial examines internal research and Meta’s response to these issues, with CEO Mark Zuckerberg testifying about platform safety and corporate priorities.
As of March 13, 2026, Meta has delayed the launch of its new AI model, Avocado, to May after internal tests showed it underperformed compared to Google's latest Gemini 3.0. Meanwhile, Meta acquired Moltbook, a social platform for AI agents, integrating its founders into Meta's AI research division to advance AI agent technology.
The 2026 World Happiness Report reveals a significant decline in life satisfaction among under-25s in Western countries, especially in the US, Canada, Australia, and New Zealand, linked to heavy social media use. Finland remains the happiest nation for the ninth year, while Costa Rica rises to fourth place. The report highlights social media's harmful effects on youth mental health, particularly among teenage girls.
In late March and early April 2026, juries in New Mexico and California found Meta liable for harming children through addictive platform design and failure to protect against sexual exploitation. Meta was ordered to pay $375 million in New Mexico and $4.2 million in California, alongside Google’s $1.8 million penalty. These landmark rulings challenge legal protections like Section 230 and signal a shift toward greater accountability for social media companies.
On March 24, 2026, a New Mexico jury found Meta liable for violating state consumer protection laws by failing to protect children on its platforms, ordering $375 million in penalties. The verdict follows a six-week trial and an undercover investigation revealing Meta concealed risks of child exploitation. Meta plans to appeal; a second trial phase in May may impose further penalties and platform changes.
The UK government is considering banning or restricting addictive features like infinite scrolling and streaks on social media platforms, following a US court ruling against Meta and Google for causing social media addiction. The consultation involves teenagers testing app limits, with potential for new laws to protect young users.
Australia's new law requires 10 major social media platforms to block under-16s or face fines of up to AU$49.5 million. While platforms have deactivated 5 million accounts, concerns remain over compliance and effectiveness, with investigations ongoing into potential non-compliance by major companies like Facebook, TikTok, and YouTube.
Australia's laws banning under-16s from major social media platforms are under scrutiny, with many children still online. Indonesia has also enforced age verification laws, accusing platforms like YouTube and Facebook of non-compliance. Both countries face challenges in enforcement and compliance from tech giants.
The UK government has been engaging social media companies to improve online safety for children. A consultation has received nearly 50,000 responses, with ongoing discussions about potential restrictions, including an Australia-style ban for under-16s. The government is considering measures to limit addictive features and AI chatbots, with decisions expected soon.