Key Takeaways New UK Online Safety Act has layers: The UK’s newest provisions to the Online Safety Act claim to protect children from harmful online content (pornographic, specifically), but is that really all there is to it? The people have spoken: Proton VPN reported a 1,400% surge in UK signups only minutes after the bill was passed, showing exactly what many people thought of the new legislation. The ‘porn block’ might involves more control than protection: The new provisions may potentially open the door to content over-moderation, privacy infringement, and security vulnerabilities. After all, online platforms (like PornHub, which suffered multiple data leaks) will have more of your personal data (like your ID). If Orwell had a social media account, it’s pretty likely he would have been shadowbanned by now. Not for being rude, but for being a little too on point. The UK has officially passed and signed into law the Online Safety Act’s newest provision At first glance, it sounds like a well-intended no-brainer: a sweeping effort to protect children from the myriad of harmful content on the internet. But dig a little deeper, and it starts to look a little unsettling. Critics argue that the law might clean up some of the internet’s darker corners, but on the flipside, it runs a high risk of turning the internet into a tightly policed echo chamber. Vague terms like ‘legal but harmful’ give platforms broad leeway to take down anything that could plausibly upset anyone. Is this law a genuine attempt to make the online world safer? Or is it simply a convenient vehicle to exert a little more control over it? What the Law Says, and Why People Are Worried The Online Safety Act gives Ofcom, the UK’s communications regulator (or media watchdog as referred to by the Brits), broad powers to oversee digital platforms. This ranges from social media giants like X (formerly Twitter) to messaging apps, forums, and even online games. It affects virtually all platforms that let UK users interact with one another, and this includes companies not headquartered in the UK but with a large UK user base. Here’s a summary of what changed on July 25, 2025: The ‘Are you 18?’ checkbox was replaced with an age verification process Facial age estimation and email-based age verification have become necessary steps Banks and mobile providers will be able to confirm your adult status Official ID verification (driver’s license or passport) to access ‘potentially harmful’ platforms Platforms enforce more restrictive content controls for children Online services must report on actions they take to keep children safe on their platforms Ofcom’s stated goal: to protect users, particularly children, from online harm. This includes a few things that most of us can generally agree are harmful: cracking down on child exploitation, terrorism-related content, and cyberbullying. However, it gets a little murky with the phrase ‘legal but harmful.’ This gives regulators the right to pressure platforms into removing content that isn’t illegal but that, for whatever reason, they deem harmful. This may sound harmless on the surface. But who defines what’s ‘harmful’? And to whom is it harmful exactly? Without a more rigid definition for these things, platforms can (and most likely will) over-police and pre-emptively remove anything ‘potentially’ offensive: a political meme, a protest video, or simply an awkward joke. You know that running gag about people yelling ‘I feel offended by this?’ Well, the UK might have just officialized legal repercussions for these things. This isn’t just hypothetical. Even before the law fully kicked in, users and moderators had reported increased takedowns on platforms like Discord, Reddit, and X. Entire communities have disappeared. Conversations once considered edgy or critical are being throttled, flagged, or shadowbanned. When your law relies on self-perceived emotional harm to police online content, platforms will start being afraid to host speech that isn’t illegal but that might cause someone to complain. And that’s a losing move for internet users, free speech, and digital freedom. The Public Reaction Was Instant (and Loud) Government officials framed the Online Safety Act as a protective measure; the public, however, seemed to interpret it differently. Within minutes of the law going into effect, UK residents started making moves online, and fast. Proton VPN, one of the world’s most privacy-focused VPN providers, reported a 1,400% surge in UK signups just minutes after the bill had passed. No, that’s not a typo. It was the largest spike they’d ever seen from a single country, and unlike other VPN surges triggered by one-off censorship events, such as France’s adult site block, this one didn’t fade. This reaction says a lot more than any think tank or government press release could. People saw the law for what it could become: a way to monitor, censor, and control. And, accordingly, they responded by masking up digitally and protecting themselves as best they could. There have even been over 446,000 signatures for a petition to repeal the Act. That’s almost half a million people that joined in protest in less than a week since the new provisions passed into law. VPNs, privacy forums, and encrypted apps are no longer niche. They’re becoming a necessary digital self-defense. And remember, this whole UK ‘porn block’ initially started from a 2015 poll made by OnePoll, a survey company known for polls like ‘The World’s Coolest Man Bun’ and ‘Wrong Side of the Bed: Myth or Fact.’ Was It Ever Just About Protecting the Children? On paper, the Online Safety Act is about shielding children from harmful content: a cause we can all get behind. But if that was really the goal, why are adults losing access to forums, private servers, and even something as harmless as political memes? The truth is, this law reaches far beyond explicit material. It’s already affecting: Gaming communities on Discord, where moderators and moderation bots receive directions to remove ‘harmful’ messages Political discourse on X, where posts criticizing UK policy are disappearing or receiving warnings Encrypted chats, where the threat