Monday, March 9, 2026

The real story isn’t AI replacing compliance officers. It’s AI removing the operational friction that kept compliance stuck in paperwork.

Jean-Christian Le Meur
Written by
Jean-Christian Le Meur
7 min read
The real story isn’t AI replacing compliance officers.  It’s AI removing the operational friction that kept compliance stuck in paperwork.

Anthropic’s new research on AI and jobs is a useful reality check for compliance leaders, not because it predicts mass layoffs, but because it measures something most debates ignore: the gap between what AI could do and what people actually use it for.

In Labor market impacts of AI: a new measure and early evidence (published March 5, 2026), Anthropic introduces “observed exposure,” combining theoretical LLM capability with real usage patterns (and weighting automated, work-related usage more heavily). Their core finding is simple: AI is far from its theoretical capability in real workflows. That “uncovered area” is where compliance work lives.

Source: Anthropic research report

What the Anthropic report changes for legal and compliance teams

A lot of commentary treats AI adoption as a matter of individual motivation (curious professionals adopt, conservative ones resist). The report points in a different direction.

It shows a persistent, measurable distance between:

🔵 what AI could already do in a job🔴 what people are actually using it for

And in Legal, the gap is massive.

IMG_2657

For compliance teams, the takeaway is not “LLMs are weak.” It is: general-purpose chat interfaces do not match systematic, evidentiary work.

2 structural blind spots that keep AI adoption low in compliance

The adoption gap in legal and compliance is often misdiagnosed as cultural. In practice, two structural blind spots explain most of it.

Blind spot 1: teams buy chatbots, but compliance runs on systems

Ask a question, get an answer, is a thin slice of compliance.

A defensible compliance program under frameworks such as France’s loi Sapin II (and AFA expectations), ISO 37001, UNE 19601, or UNE 19603 is built on artifacts you can revisit years later:

  • risk mapping by activity, geography, and third parties
  • translation of non-compliance risks into operational controls
  • evidence collection, verification, and retention
  • traceability (who did what, when, based on what)
  • repeatable testing of control effectiveness, not just control design

Most of the workflows compliance teams rely on today were designed long before AI was part of the picture.

A chat interface can draft a policy clause. It cannot, on its own, maintain a risk-control-evidence chain.

If the tool cannot connect AI output to your compliance operating model, adoption stalls because using it creates extra work (copy-paste, reformatting, revalidating).

Blind spot 2: in regulated environments, trust is a design requirement

In many roles, a wrong answer is a productivity issue.

In compliance, a wrong answer can become:

  • a failed audit (internal, external, certification)
  • a regulatory challenge on program effectiveness
  • reputational damage
  • executive exposure

So the real question becomes: could you defend the AI-assisted decision to a regulator or auditor in three years?

That pushes you toward AI that is:

  • traceable in sources (or clearly limited when it cannot cite)
  • auditable in outputs (versioning, retention, approvals)
  • governed (who can use it, for what, with what checks)
  • embedded in your processes, not parallel to them

7 blind spots blocking AI in compliance, and how to build audit-ready evidence

These are the adoption blockers that show up in real programs (France and Spain included). Each comes with a concrete fix you can implement without changing your whole stack.

1) Using AI for “answers” instead of for evidence-heavy workflows

Symptom: the team tests AI on legal interpretation questions, gets nervous about hallucinations, and stops.

Fix: start where AI reduces friction while keeping humans in charge.

Best first workflows are typically:

  • obligation extraction into a structured register (with human validation)
  • drafting first versions of policies, procedures, control narratives
  • converting scattered evidence into indexed, searchable records
  • preparing audit packs (assembling what already exists, highlighting gaps)

The win is not “AI decided.” The win is “AI reduced the time to assemble and verify.”

2) No separation between control design and control effectiveness testing

Symptom: AI drafts controls and the organization treats that as “done.”

Fix: keep a strict split:

  • design: what the control is supposed to do (policy, workflow, threshold, approval)
  • testing: how you prove it worked (sampling rules, evidence criteria, exceptions, remediation)

AI can help with both, but your governance must label which is which, and store evidence accordingly.

3) No “minimum defensibility standard” for AI-assisted work

Symptom: AI is used inconsistently, sometimes with no record of prompts, inputs, or reviewer.

Fix: define a minimum evidence record for any AI-assisted compliance artifact.

If you cannot capture this lightly, AI will remain “off the record,” and leadership will (rationally) resist scaling it.

4) Risk mapping remains a static spreadsheet, so AI has nothing stable to anchor to

Symptom: you run a risk assessment workshop, update a heatmap once a year, and the rest of the year everyone works from memory.

Fix: convert the risk map into a living object with three stable identifiers:

  • a risk ID and scenario definition
  • linked controls (preventive/detective, owner, cadence)
  • linked tests and evidence

This is also where cross-border complexity is easiest to manage: you can align a core scenario library, then add France- and Spain-specific overlays (Sapin II specifics, UNE 19601 offence categories, UNE 19603 antitrust scenarios) without duplicating everything.

5) Third-party due diligence is treated as intake paperwork, not an owned operating process

Symptom: due diligence happens at onboarding, then quietly decays. Evidence is scattered across procurement, legal, and finance.

Fix: run third-party oversight as a lifecycle with operational ownership:

  • onboarding risk tier and required controls
  • contract clauses mapped to controls (not just “legal terms”)
  • periodic re-review triggers
  • remediation tracking for third parties (not just internal actions)

A concrete example: if marketing wants to use geo-distributed social publishing services such as TokPortal for posting TikToks in target countries, treat it like any other third party. Define what you need to evidence (ownership, access control, data handling, brand/account governance), who approves, and how you re-check it.

6) Training fatigue is solved with “more content” instead of measurable behavior

Symptom: training completion rates exist, but you cannot show whether behavior changed.

Fix: define 2 to 3 outcome signals per training topic (and capture them as evidence). Examples:

  • gifts and hospitality: approval rate vs. threshold breaches, register completeness
  • antitrust: pre-clearance usage for trade association meetings, minutes checks
  • speak-up: time to triage, substantiation rate, feedback timeliness

AI can help draft micro-scenarios, localize content, and generate manager discussion guides, but your program must measure outcomes or auditors will see “activity, not effectiveness.”

7) Reporting to leadership stays narrative-only, so resourcing stays political

Symptom: the board hears “we updated policies,” not “risk decreased and here is the proof.”

Fix: adopt a small set of evidence-backed KPIs that connect directly to program effectiveness. Even before dashboards, define consistent rules for:

  • control testing coverage (what percent of key controls were actually tested)
  • exception rate and remediation velocity
  • third-party coverage by tier
  • audit evidence retrieval time (a simple but revealing metric)

Anthropic’s “observed exposure” idea is useful here: you can measure your own “observed automation” in compliance. Not to cut headcount, but to show increased capacity and reduced operational drag.

This is how you avoid the most common trap: scaling AI because it demos well, then discovering it cannot be defended.

A simple diagram showing two overlapping areas: theoretical AI capability (larger circle) and observed usage (smaller circle), with a highlighted “adoption gap” zone labeled “workflow friction + trust requirements” and a small note “compliance work lives here.”

Frequently asked questions

Why is AI adoption slower in legal and compliance than in other knowledge work? Because compliance is evidentiary and must be defensible over time. Tools that only provide conversational answers do not produce the traceability, workflow, approvals, and retention required for audits and regulators.

What is the safest place to start with AI in a Sapin II or ISO 37001 program? Start with friction removal where humans still decide: evidence collection, obligation extraction into registers, first drafts of policies and control narratives, and audit pack assembly with clear reviewer sign-off.

How do we make AI outputs auditable? Treat AI-assisted artifacts like any other controlled document: capture inputs, the AI operation (prompt/tool), outputs, human review, approval, versioning, and retention location. If you cannot reproduce or explain it, do not scale it.

How naltilia can help

Naltilia is designed for the “adoption gap” highlighted by Anthropic: not AI as a standalone chat, but AI embedded into compliance workflows. It helps teams operationalize regulatory risk assessment, translate risks into owned remediation actions, automate data and evidence collection, and run workflow automation with traceability. The goal is to reduce paperwork friction while improving audit readiness across programs aligned to loi Sapin II, ISO 37001, and UNE 19601/19603.

Contact Naltilia to discuss your workflow and evidence priorities: https://calendly.com/iratxe-naltilia/30min

This article is general information, not legal advice.

About the Author

Jean-Christian Le Meur

Jean-Christian Le Meur

I'm an AI startup founder and CEO with years of executive experience in large, complex organizations. My passion lies in helping companies unlock tangible business value through data and emerging technologies, always with a human-centric approach. I thrive where strategy meets execution, leading teams to deliver results across international markets. For me, technology is not the destination, it's the tool. The real transformation happens when people and technology work together to create lasting impact.