
Naomie Halioua
Co-founder & CRO, AI Research

What If Someone Fed Your Compliance AI a Fake GDPR Text?
Every week, I read dozens of research papers on AI and regulatory compliance to select just one β the most useful, the most actionable, the one that truly changes how you think about the subject. This week, one question stopped me: what happens when the regulatory text your AI is reading has been deliberately falsified? And does your system even notice?
Safer Policy Compliance with Dynamic Epistemic Fallback (DEF)
Imperial, Tayyar Madabushi Β· arXiv:2601.23094, January 2026
"Humans develop a series of cognitive defenses, known as epistemic vigilance, to combat risks of deception and misinformation. Developing safeguards for LLMs inspired by this mechanism might be particularly helpful for their application in high-stakes tasks such as automating compliance with data privacy laws."
Imperial & Tayyar Madabushi, arXiv:2601.23094
What the paper introduces
The paper introduces DEF, a protocol that gives LLMs the equivalent of human epistemic vigilance. When the system encounters a regulatory text that has been maliciously modified, it flags the inconsistency, refuses to comply, and falls back to its own verified knowledge. Tested on GDPR and HIPAA, with DeepSeek-R1 reaching 100% detection in one setting.
100%
detection (DeepSeek-R1)
GDPR
& HIPAA tested
DEF
first defense framework
The detail that changes everything
Most compliance AI systems assume the regulatory text they read is authentic. DEF is the first framework built on the opposite assumption β that the input might be compromised. In an environment where legal artifacts can be manipulated, that is not a theoretical risk.
A falsified GDPR excerpt fed to an unprotected LLM can produce a confidently wrong compliance output. The AI will not hesitate, it will not flag uncertainty β it will apply the fake rules as if they were real. DEF is the first documented defense against this attack vector.
Why this matters to you
You are a DPO or compliance officer
If your organization uses an AI tool to interpret or apply regulatory texts, this paper introduces a threat you may not have considered. A falsified GDPR excerpt fed to an unprotected LLM can produce a confidently wrong compliance output. DEF is the first documented defense.
You are building an AI product
The paper shows that frontier models like DeepSeek-R1 can be pushed to 100% detection with the right inference-time protocol. That is an engineering decision, not a research problem. If your compliance layer reads external policy texts, this should be on your architecture checklist.
Your sector is regulated (healthcare, finance, energy, HR)
The paper tests on GDPR and HIPAA explicitly. Legal artifacts in regulated sectors are exactly the type of document that adversarial actors would target. If your compliance workflow relies on AI reading regulatory texts, the integrity of those texts is a security concern.
How Cleo handles this
At Cleo, every regulatory text we process goes through a verification layer before it enters our pipeline. This paper formalizes what we treat as a non-negotiable: the integrity of regulatory inputs is a prerequisite, not an afterthought. Our 30+ specialized agents cross-validate sources against official databases, flag inconsistencies, and escalate to human review when confidence is low.
Reference: Imperial & Tayyar Madabushi (2026), Safer Policy Compliance with Dynamic Epistemic Fallback, arXiv:2601.23094
Frequently asked questions
What is epistemic vigilance in AI compliance?
Epistemic vigilance is a concept borrowed from cognitive science: humans develop cognitive defenses to detect deception and misinformation. Applied to AI compliance, it means building systems that can detect when a regulatory text has been falsified or tampered with, rather than blindly following any input labeled as law.
What is the DEF framework for compliance AI?
DEF (Dynamic Epistemic Fallback) is a protocol introduced by Imperial & Tayyar Madabushi (arXiv:2601.23094, January 2026). When an LLM encounters a regulatory text that has been maliciously modified, DEF flags the inconsistency, refuses to comply with the falsified text, and falls back to verified knowledge. Tested on GDPR and HIPAA, DeepSeek-R1 reached 100% detection in one setting.
Can someone really fake a GDPR text to trick an AI?
Yes. Most compliance AI systems ingest regulatory texts from external sources without verifying their authenticity. A falsified GDPR excerpt β with subtly altered obligations, removed safeguards, or fabricated exceptions β can produce confidently wrong compliance output from an unprotected LLM. This is not a theoretical risk: adversarial document injection is a known attack vector for retrieval-augmented generation (RAG) pipelines.
Related resources
Solutions
Product Compliance SolutionTry Cleo: free regulatory risk scan
See your regulatory landscape mapped in minutes. No signup, no credit card.