OpenAI and Google are about to have a massive legal problem. It affects YOUR A.I. – BAM – Whatfinger News' General Dispatch
m
Recent Posts
Connect with:
Thursday / April 2.
HomeWhatfinger NewsOpenAI and Google are about to have a massive legal problem. It affects YOUR A.I. – BAM

OpenAI and Google are about to have a massive legal problem. It affects YOUR A.I. – BAM

OpenAI, Google, and Anthropic have repeatedly sworn to courts that their models do not store exact copies of copyrighted books. They claim their “safety training” prevents regurgitation. Researchers just dropped a paper called “Alignment Whack-a-Mole” that proves otherwise. They didn’t use complex jailbreaks or malicious prompts. They just took GPT-4o, Gemini, and DeepSeek, and fine-tuned them on a normal, benign task: expanding plot summaries into full text. The safety guardrails instantly collapsed. Without ever seeing the actual book text in the prompt, the models started spitting out exact, verbatim copies of copyrighted books. Up to 90% of entire novels, word-for-word. Continuous passages exceeding 460 words at a time.

But here is the part that changes everything. They fine-tuned a model exclusively on Haruki Murakami novels. It didn’t just learn Murakami. It unlocked the verbatim text of over 30 completely unrelated authors across different genres. The AI wasn’t learning the text during fine-tuning. The text was already permanently trapped inside its weights from pre-training. The fine-tuning just turned off the filter. It gets worse. They tested models from three completely different tech giants. All three had memorized the exact same books, in the exact same spots. A 90% overlap. It’s a fundamental, industry-wide vulnerability. For years, AI companies have argued in court that their models are just “learning patterns,” not storing raw data. This paper provides the smoking gun.

  • GPT-4o, Gemini, and DeepSeek all memorized the same books in the same spots. 90% overlap across three competitors. They didn’t train together. So where did they all get the same data?
  • AI companies have told courts for years: Our models don’t store copies alignment prevents regurgitation Alignment Whack-a-Mole just showed that a normal writing-assistant fine-tune (expand plot summary) bypasses every guardrail. Result: 85-90% verbatim recall of novels, 460+ word continuous passages, and cross-author unlocking (Murakami fine-tune → 30+ other authors). The memorization was already latent in the weights. Fine-tuning simply turned off the filter. This isn’t a jailbreak it’s a feature of how these models actually work. – SH

Come to Whatfinger news’ Homepage – for MORE news you want than any other site on the net – CLICK HERE





No comments

leave a comment