Legal Innovation
February 09
By alex dantart
For months we repeated a calming mantra: "AI will not replace you, someone who uses it will." Today I have to express something uncomfortable: That phrase is now obsolete..
And it becomes even clearer after analyzing an empirical study with 2.700 real cases that demonstrates something many of us suspected but no one had measured: Simply using AI is not enough. We need to know what kind of AI to use, or the cure will be worse than the disease..
The experiment that changes everything
Twelve of the best models on the market (GPT-5, Claude, Gemini, etc.) were evaluated by drafting 75 real legal tasks: appeals, objections to precautionary measures, and jurisprudential reasoning. Three different scenarios were used:
The results are devastating, and could give more than one person sleepless nights. When a lawyer asks ChatGPT or Claude to draft an appeal "from memory," almost 3 out of every 10 quotes it generates are fakeNon-existent rulings, incorrect attributions, fabricated legal doctrine. A 26.8% error rate in citations and a 15.6% error rate in fabricated facts.
Does that sound like an exaggeration? Perhaps not so much for that lawyer sanctioned by the Constitutional Court for submitting 19 nonexistent citations. Or for the Italian and Argentinian colleagues recently reprimanded for the same thing.
But there's more: each document generated in this way It requires an average of 35 minutes of review time. to correct it, so it's not a useful draft, but an information liability that wastes more time than it saves.
Here's the twist: when those same AIs work based on verified sources (which is achieved with RAG technology), the error rate drops to 8.3%. And with advanced verification systems, practically disappears: 0.046%Furthermore, the review time is reduced from 35 minutes to… 1.2 minutes.
The difference isn't in "using more AI." It's in understanding that there are two radically different types of legal AI:
The creative oracle (pure generative AI) that:
The expert archivist (consultative AI) who:
Let's return to the initial idea: "You'll be replaced by someone using AI." Well, no. Because By the end of the year we'll probably all be using itWord will incorporate AI as standard, legal databases will integrate it, courts will implement it, and legaltech solutions thrive on it. The real divide won't be between those who use AI and those who don't. It will be between:
Three questions we should ask ourselves today
This study demonstrates something revealing: Digital competence is no longer enough. Need architectural competitionKnowing when we need creativity and when we need absolute rigor. It's not about being afraid of AI. It's about being afraid of misuse.
Because when the Constitutional Court issues a ruling, it doesn't ask what model you used. It asks why it wasn't verified. And "It was generated by AI" It is not a defense but rather an aggravating factor.
The profession will not be decided by who uses AI, but by who understands what type of AI is needed at any given time.
And in whom does he preserve something that no machine can replicate: the ability to put their name and responsibility behind every decision.
Note: The complete "Reliability by Design" study with the 2.700 cases analyzed is available at arxiv.orgIt includes the JURIDICO-FCR dataset for replication.
(Link: https://arxiv.org/abs/2601.15476)