This website uses cookies

Read our Privacy policy and Terms of use for more information.

The bot economy didn't break the internet by accident — it broke it because no governance framework existed to stop it. This week, we move past the diagnosis and into the harder question: what would it actually take to fix an AI ecosystem built on a decade of poisoned data?

The Problem in One Paragraph

If you haven't watched this week's video yet — go watch it first. Link below.

Nearly half of all internet traffic is generated by bots. AI security systems trained on that traffic learned fake behavior as their baseline for "human." The bots then studied the detection models and evolved to evade them. The result is an internet — and an AI ecosystem — built on a foundation that was never fully real. The platforms profited. The regulators arrived late. And nobody was ever structurally required to tell the truth about the numbers.

That ends only if governance makes it unavoidable.

Why Governance Has Failed So Far

Before prescribing solutions, we have to be honest about why existing frameworks have consistently underperformed:

  • Self-reporting incentives are backwards. Platforms profit from high engagement numbers — real or fake. Requiring them to self-audit bot traffic is like asking a restaurant to grade its own health inspection. The conflict of interest is structural, not incidental.

  • Regulation has chased the wrong metric. Most platform legislation focuses on content moderation — what gets said. Almost none addresses engagement integrity — whether the signals amplifying that content are real. GDPR, DSA, and the EU AI Act all largely sidestep this.

  • AI liability is still unassigned. When an AI model makes a consequential decision based on training data that was contaminated by synthetic activity, there is currently no legal framework in any jurisdiction that assigns clear accountability for that outcome.

  • The speed gap is fatal. By the time a regulatory framework is finalized, bot networks have already reverse-engineered the detection models it relies on. Compliance cycles measured in years cannot keep up with adversarial systems that adapt in weeks.

What Effective Governance Actually Looks Like

This is the section your regulators need to read. These aren't hypothetical ideals — they are technically implementable, legally precedented, and politically achievable with sufficient will.

1. Mandatory Engagement Integrity Audits

Platforms above 10 million monthly active users should be required to submit quarterly independent audits of traffic authenticity — conducted by third-party firms with no platform revenue relationship. Results must be published in standardized formats, comparable across platforms. Think financial auditing, applied to engagement data. The SEC already requires this level of scrutiny for revenue figures. Engagement numbers that drive ad pricing — and AI training datasets — deserve the same standard.

2. AI Training Data Provenance Certification

Any AI model trained substantially on web-scraped data and deployed in high-stakes contexts (finance, healthcare, law enforcement, elections) should be required to carry a data provenance certificate — a verifiable disclosure of what percentage of training interactions were independently confirmed as human-generated. This isn't technically impossible; it's just currently optional. It should not be optional when the model is making decisions about your loan, your medical record, or your bail hearing.

3. Adversarial Mimicry as a Distinct Legal Category

Current law treats bots primarily as a fraud or spam issue. It does not specifically address bots engineered to evade AI detection systems — a materially different and more dangerous category. Adversarial mimicry should carry elevated criminal liability, equivalent to tampering with a regulatory instrument. If you build software specifically designed to fool a court-mandated security system, that is not just spam. That is obstruction.

4. Cross-Platform Bot Intelligence Sharing — With Teeth

Bot networks don't respect platform boundaries — they operate across Instagram, X, TikTok, YouTube, and Reddit simultaneously. Detection, however, is siloed per platform. Governance should mandate a shared threat intelligence framework, similar to how financial institutions are required to report suspicious activity to FinCEN. Platforms should be legally required to share identified bot behavioral signatures with a central registry — accessible to all platforms and independent researchers — within 72 hours of detection.

5. A "Synthetic Traffic" Disclosure Standard for Advertisers

Advertisers are paying real money for fake eyeballs, and most have no legal recourse when they find out. A mandatory synthetic traffic disclosure standard — requiring platforms to notify advertisers when post-campaign audits find that a defined threshold of paid impressions were non-human — would create immediate financial accountability. When platform revenue is directly tied to traffic authenticity, the incentive structure finally points in the right direction.

6. Independent AI Baseline Integrity Testing

Governance on this specific issue is embryonic — but not entirely absent:

Jurisdiction

Current Status

EU

Digital Services Act requires "very large platforms" to assess systemic risks — but bot baseline contamination is not explicitly named

USA

BOT Act (2023 proposal) targets social media bots for political ads only — scope is far too narrow

UK

Online Safety Act focuses on harmful content, not engagement integrity

Canada

No specific legislation — AIDA (proposed AI Act) does not address training data contamination

China

Requires real-name registration, which reduces some bot activity — but enforcement serves state interests, not independent oversight

The honest assessment: no jurisdiction has yet passed legislation that directly addresses AI training data contamination from synthetic web traffic. This is a wide-open governance gap.

What the Research Community Needs

Governance can't outpace a problem it can't measure. Policymakers should fund:

  • Longitudinal bot traffic studies with open datasets — not dependent on platform cooperation

  • Red team programs that specifically test AI security systems for adversarial mimicry vulnerabilities, with published results

  • Academic independence protections — formal whistleblower-style protections for researchers whose findings inconvenience large platforms

The researchers who surfaced the Dead Internet problem did so without institutional support and faced quiet professional consequences. That is not a sustainable model for a problem of this scale.

What You Can Do as an Informed Citizen

Governance moves faster when constituents understand the issue at stake. Three concrete actions:

  1. Ask your representative whether your country's AI legislation addresses training data provenance — most don't know, and the question itself creates accountability

  2. Support independent researchers covering platform integrity — subscribe, cite, amplify their work without the filter of platform algorithms

  3. When evaluating AI tools used in high-stakes decisions affecting you — ask the vendor directly: What was this model trained on, and how was the integrity of that data verified? You are entitled to an answer

Watch This Week's Full Video First

The newsletter gives you the solutions. The video gives you the full story of how we got here — and why the scale of the problem is bigger than most people realize.

Keep Reading