June 23, 2025
The GIST The law relies on being precise. AI is disrupting that
Lisa Lock
scientific editor
Andrew Zinin
lead editor
Editors' notes
This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:
fact-checked
trusted source
proofread

The use of AI is spreading throughout all professions and industries—and law is no exception.
Lawyers have been innovative in their use of large language models (LLMs), but we're also seeing examples of them being caught out by the language capabilities of generative AI.
Don't rely on AI
Refined AI models, combined with retrieval or "RAG" systems, have the capacity to summarize, review and analyze documents.
But increasingly, courts and legal regulators are warning lawyers not to rely on generative AI without checking their work—and in some instances are advising them not to use it at all.
In fact, we've already seen courts admonish lawyers for submitting legal documents that include content fabricated by AI.
And in Australia, courts themselves have been cautious about the use of the technology.
Ordinary meaning and generative AI
Some courts around the world are experimenting with using generative AI.
In England, Lord Justice Birss advised that he used ChatGPT to summarize an area of the law. In the U.S. case of Snell v United Specialty Insurance Co, Justice Kevin Newsom used ChatGPT to determine the plain and ordinary meaning of a contentious term in an insurance policy.
When interpreting documents to decide their legal meaning, courts typically look to the ordinary meaning of words, as well as the context in which they are used.
But how do judges determine the ordinary meaning of a word?
One approach is to ask "ordinary people."
For most of the 19th century, the U.S. Supreme Court mandated that its justices ride town-to-town hearing cases to give them exposure to everyday citizens and conditions outside the capital.
Judges often consult a dictionary. Or perhaps they ask ChatGPT.
One tool among many
Let's go back to the Snell case in the U.S. mentioned earlier. In mid-2024, Justice Kevin Newsom was tasked with deciding whether an insurance policy's coverage of "landscaping" included the installation of a trampoline.
Justice Newsom checked three dictionaries and found three very different answers.
His Honor considered a "visceral, gut-instinct" feeling, only to decide it didn't seem very legally compelling. Instead, His Honor asked ChatGPT.
The possibility of generative AI providing an "ordinary meaning" is in some ways compelling.
After all, these models are trained on vast corpuses of the English language—books, newspapers, user prompts—covering all sorts of speech in all sorts of contexts.
Their reading is not limited by background, interests or age. In this sense, ChatGPT might be seen as representing an amalgam of the ordinary person in a way that judges, or even the compilers of dictionaries, are not.
All that said, the judge ended by sounding a note of caution. In his Honor's view, LLMs should only be "one tool among many," held up and tested against historical context, common sense and dictionary meanings.
Not an oracle of ordinary meaning
There are also important differences between generative AI and dictionaries, which limit their role as oracles of ordinary meaning.
As numerous lawyers have discovered, not only does generative AI hallucinate, it can also be sycophantic, providing the answer that, as suggested by the context, is what the person providing the prompt wants to hear.
This means the tools are likely to offer up confident-but-rubbish answers by attempting to mimic human speech.
Another concern is around transparency.
We know how dictionaries are compiled and the process is scrupulously documented. We have no clear understanding of the training data used in the free general-purpose AI models.
And, unlike dictionaries, which use a collaborative process for compiling content, much of the AI training data is often used without the knowledge or consent of the original author.
From common meaning to judgment
Courts, businesses, governments and firms are continuing to experiment with generative AI. Tech firms and developers are continuing to find ways to refine and improve their outputs.
For lawyers and courts, accuracy matters. But so do other values, like transparency, fairness and accountability.
These are part of Australia's ethical AI framework and central to the administration of justice.
A judge's core responsibility in interpreting texts has remained unchanged across the centuries: resolving the tension between clinical legal language and the often-messy nature of our lived realities.
The allure of generative AI's data-backed outputs should not distract from the fact that these decisions invariably rely on judgment and context.
While it may present new pieces of information in making these decisions, generative AI cannot be treated as any more authoritative or reliable than any other source—and certainly no more ethically compelling.
Provided by University of Melbourne Citation: The law relies on being precise. AI is disrupting that (2025, June 23) retrieved 23 June 2025 from https://techxplore.com/news/2025-06-law-precise-ai-disrupting.html This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
Explore further
Justice at stake as generative AI enters the courtroom shares
Feedback to editors