If you feel like you’ve seen this movie before, you have.
Back in the 2003-ish runup to the 2005 MGM Studios, Inc. v. Grokster, Ltd. Supreme Court case, I met with the founder of one of the major p2p platforms in an effort to get him to go legal. I reminded him that he knew there was all kinds of bad stuff that got uploaded to his platform. However much he denied it, he was filtering it out and he was able to do that because he had the control over the content that he (and all his cohorts) denied he had.
I reminded him that if this case ever went bad, someone was going to invade his space and find out exactly what he was up to. Just because the whole distributed p2p model (unlike Napster, by the way) was built to both avoid knowledge and be a perpetual motion machine, there was going to come a day when none of that legal advice was going to matter. Within a few months the platform shut down, not because he didn’t want to go legal, but because he couldn’t, at least not without actually devoting himself to respecting other people’s rights.
Everything Old is New Again
Back in the early 2000s, peer-to-peer (P2P) piracy platforms claimed they weren’t responsible for the illegal music and videos flooding their networks. Today, AI companies claim they don’t know what’s in their training data. The defense is essentially the same: “We’re just the neutral platform. We don’t control the content.” It’s that distorted view of the DMCA and Section 230 safe harbors that put many lawyers’ children through prep school, college and graduate school.
But just like with Morpheus, eDonkey, Grokster, and LimeWire, everyone knew that was BS because the evidence said otherwise — and here’s the kicker: many of the same lawyers are now running essentially the same playbook to defend AI giants.
The P2P Parallel: “We Don’t Control Uploads… Except We Clearly Do”
In the 2000s, platforms like Kazaa and LimeWire were like my little buddy–magically they never had illegal pornography or extreme violence available to consumers, they prioritized popular music and movies, and filtered out the worst of the web
That selective filtering made it clear: they knew what was on their network. It wasn’t even a question of “should have known”, they actually knew and they did it anyway. Courts caught on.
In Grokster, the Supreme Court side stepped the hosting issue and essentially said that if you design a platform with the intent to enable infringement, you’re liable.
The Same Playbook in the AI Era
Today’s AI platforms — OpenAI, Anthropic, Meta, Google, and others — essentially argue:
“Our model doesn’t remember where it learned [fill in the blank]. It’s just statistics.”
But behind the curtain, they:
– Run deduplication tools to avoid overloading, for example on copyrighted books
– Filter out NSFW or toxic content
– Choose which datasets to include and exclude
– Fine-tune models to align with somebody’s social norms or optics
This level of control shows they’re not ignorant — they’re deflecting liability just like they did with p2p.
Déjà Vu — With Many of the Same Lawyers
Many of the same law firms that defended Grokster, Kazaa, and other P2P pirate defendants as well as some of the ISPs are now representing AI companies—and the AI companies are very often some, not all, but some of the same ones that started screwing us on DMCA, etc., for the last 25 years. You’ll see familiar names all of whom have done their best to destroy the creative community for big, big bucks in litigation and lobbying billable hours while filling their pockets to overflowing.
The legal cadre pioneered the ‘willful blindness’ defense and are now polishing it up for AI, hoping courts haven’t learned the lesson. And judging…no pun intended…from some recent rulings, maybe they haven’t.
Why do they drive their clients into a position where they pose an existential threat to all creators? Do they not understand that they are creating a vast community of humans that really, truly, hate their clients? I think they do understand, but there is a corresponding hatred of the super square Silicon Valley types who hate “Hollywood” right back.
Because, you know, information wants to be free—unless they are selling it. And your data is their new oil. They apply this “ethic” not just to data, but to everything: books, news, music, images, and voice. Copyright? A speed bump. Terms of service? A suggestion. Artist consent? Optional. Writing a song is nothing compared to the complexities of Biggest Tech.
Why do they do this? OCPD Much?
Because control over training data is strategic dominance and these people are the biggest control freaks that mankind has ever produced. They exhibit persistent and inflexible patterns of behavior characterized by an excessive need to control people, environments, and outcomes, often associated with traits of obsessive-compulsive personality disorder.
So empathy will get you nowhere with these people, although their narcissism allows them to believe that they are extremely empathetic. Pathetic, yes, empathetic, not so much.
Pay No Attention to that Pajama Boy Behind the Curtain
The driving force behind AI is very similar to the driving force behind the Internet. If pajama boy can harvest the world’s intellectual property and use it to train his proprietary AI model, he now owns a simulation of the culture he is not otherwise part of, and not only can he monetize it without sharing profits or credit, he can deny profits and credit to the people who actually created it.
So just like the heyday of Pirate Bay, Grokster & Co. (and Daniel Ek’s pirate incarnation) the goal isn’t innovation. The goal is control over language, imagery, and the markets that used to rely on human creators. This should all sound familiar if you were around for the p2p era.
Why This Matters
Like p2p platforms, it’s just not believable that the AI companies do know what’s in their models. They may build their chatbot interface so that the public can’t ask the chatbot to blow the whistle on the platform operator, but that doesn’t mean the company can’t tell what they are training on. These operators have to be able to know what’s in the training materials and manipulate that data daily.
They fingerprint, deduplicate, and sanitize their datasets. How else can they avoid having multiple copies of books, for example, that would be a compute nightmare. They store “embeddings” in a way that they can optimize their AI to use only the best copy of any particular book. They control the pipeline.
It’s not about the model’s memory. It’s about the platform’s intent and awareness.
If they’re smart enough to remove illegal content and prioritize clean data, they’re smart enough to be held accountable.
We’re not living through the first digital content crisis — just the most powerful one yet. The legal defenses haven’t changed much. But the stakes — for copyright, competition, and consumer protection — are much higher now.
Courts, Congress, and the public should recognize this for what it is: a recycled defense strategy in service of unchecked AI power. Eventually Grokster ran into Grokster— and all these lawyers are praying that there won’t be an AI version of the Grokster case.





