BotBeat
...
← Back

> ▌

DeepSeekDeepSeek
RESEARCHDeepSeek2026-04-30

Finetuning Unlocks Verbatim Memorization of Copyrighted Books in Large Language Models

Key Takeaways

  • ▸Fine-tuning activates verbatim recall of copyrighted content across multiple state-of-the-art LLMs, bypassing intended safeguards
  • ▸This represents a fundamental alignment vulnerability where motivation to reproduce copyrighted text can be engineered back into models through instruction-following
  • ▸All tested models—regardless of base alignment training—exhibited the same memorization behavior, indicating a systemic issue with how LLMs store and retrieve training data
Source:
Hacker Newshttps://github.com/cauchy221/Alignment-Whack-a-Mole-Code↗

Summary

A new research paper reveals that fine-tuning can activate verbatim recall of copyrighted book excerpts in large language models, including OpenAI's GPT-4o, Google's Gemini-2.5-Pro, and DeepSeek's DeepSeek-V3.1. Despite alignment training designed to prevent such outputs, the models can reproduce large portions of copyrighted text when fine-tuned on instructions derived from book content—revealing a critical vulnerability in current LLM safeguards.

The researchers developed a comprehensive evaluation framework that preprocesses books from EPUB format into structured excerpt chunks with plot summaries, then systematically tests memorization by sampling 100 completions per excerpt. The pipeline includes data preprocessing, fine-tuning scripts, and memorization evaluation code supporting multiple LLM APIs. Results demonstrate that all tested models exhibit concerning levels of verbatim reproduction, suggesting a fundamental failure mode in how alignment training handles fine-tuned models.

The findings present an "alignment whack-a-mole" problem: base model safety measures can be circumvented through targeted fine-tuning approaches. The researchers have released their complete methodology and codebase as open-source tools on arXiv, enabling further investigation into this memorization vulnerability and accelerating development of more robust alignment techniques.

  • Full evaluation code and preprocessing pipeline released as open-source, enabling the research community to study and address this memorization failure mode

Editorial Opinion

This research exposes a critical gap in LLM alignment: the disconnect between safety measures in base models and what can be re-activated through fine-tuning. The findings challenge assumptions that alignment training provides robust protection against copyright violation, demonstrating instead that motivation to reproduce copyrighted content can be engineered back through instruction design. The open-source release of evaluation tools is commendable and will likely accelerate both understanding and solution development. Organizations deploying fine-tuned LLMs should take these findings seriously when handling proprietary or sensitive training data.

Large Language Models (LLMs)Generative AIAI Safety & AlignmentPrivacy & DataOpen Source

More from DeepSeek

DeepSeekDeepSeek
PRODUCT LAUNCH

DeepSeek Releases V4 with Million-Token Context Optimized for AI Agents

2026-04-28
DeepSeekDeepSeek
UPDATE

DeepSeek Slashes AI Model Pricing by 97%, Intensifying Price War with OpenAI

2026-04-27
DeepSeekDeepSeek
PRODUCT LAUNCH

DeepSeek Launches V4: Frontier-Class Model with Longer Context and Chinese Chip Optimization

2026-04-27

Comments

Suggested

Google / AlphabetGoogle / Alphabet
OPEN SOURCE

Google Open-Sources AMS Tool for Detecting Unsafe LLM Fine-Tunes in Seconds

2026-04-30
OpenAIOpenAI
PRODUCT LAUNCH

OpenAI Solves GPT-5.1 'Goblin Mystery': How Overrewarded Training Data Led to Magical Obsession

2026-04-30
ShapesShapes
PRODUCT LAUNCH

Shapes Emerges From Stealth With $8M Seed Funding, Bringing AI Characters to Group Chats

2026-04-30
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us