[CloakLLM]/Whitepaper
Technical Whitepaper · v1.0 · April 2026

The Article 12 Paradox

Why GDPR and the EU AI Act cannot both be satisfied without PII middleware.

The EU AI Act requires high-risk AI systems to log every interaction. GDPR requires personal data to be minimized and deleted. This whitepaper explains why the two obligations are structurally in conflict — and the architectural middleware layer that resolves the paradox.

~10 minute read8 sectionsMIT License · Free to share

What you'll learn

A technical, vendor-honest read for compliance officers, AI architects, and engineering leaders deploying high-risk AI systems in the EU.

1

Two laws, one conflict

Article 12 of the EU AI Act requires automatic logging of every high-risk AI interaction. GDPR Article 5 requires data minimisation and storage limitation. Both are mandatory and simultaneously in force.

2

The compliance gap

Organizations that log full interactions to satisfy the AI Act accumulate GDPR liability. Organizations that delete logs for GDPR violate the AI Act. Neither path is legal.

3

The architectural fix

Strip PII at the input layer — before it reaches any log, any model call, or any downstream system. Tokenization is GDPR-recognised pseudonymisation (Recital 26, Article 4(5)).

4

Behavioral vs. identity traceability

Article 12 requires behavioral traceability (what the system did), not identity traceability (who did it). GDPR actively prohibits the latter as a logging mechanism.

The paradox, in two sentences

EU AI Act · Article 12

High-risk AI systems must automatically record events throughout their operational lifecycle. Deployers retain logs for at least six months. Technical documentation for ten years.

GDPR · Article 5

Personal data must be minimized, purpose-limited, and deleted when the purpose is served. No exceptions carved out for AI logging.

The resolution: PII middleware

Intercept every AI interaction before it reaches the model. Strip personal data via deterministic tokenization. Pass only the sanitized form downstream — to the model, to the log, and to every other system. Logs stay complete. PII never lands in them.

What CloakLLM guarantees

Architectural properties the whitepaper walks through in detail.

  • Article 12 logs contain full operational records with zero personal data
  • GDPR data minimisation satisfied — only tokenized text and operational metadata logged
  • GDPR storage limitation satisfied — logs contain no personal data to delete
  • Tamper-evident hash chain (SHA-256) with optional Ed25519 signed certificates
  • Three-pass detection: regex, NER, and opt-in semantic LLM
  • Article 4a ready: pseudonymised special-category data for bias detection

Read it. Share it. Ship compliance-ready AI.

The whitepaper is free, MIT-licensed, and written for technical audiences — not marketing. Download the PDF or start integrating the open-source SDK today.