Series
Blogs
Series
Blogs
Artificial intelligence tools are playing an increasingly important role in the delivery of legal services. However, these tools create considerable risks. We look at some hard statistics on the predisposition of AI tools to make up cases and examine the Divisional Court’s judgment in Ayinde as an example of why this can be a real problem in practice.
The simple answer to this is yes. Large Language Models are stochastic parrots trained to create text based on probabilistic predictions. This process of regurgitating and reinventing the internet (and other learned text) creates an inherent risk that case citations will not be replicated accurately or will simply be fabricated.
The LinksAI English law benchmark gives an idea of the prevalence of this issue. The benchmark was created to test the ability of Large Language Models to answer legal questions. The first benchmark was released in October 2023 and the second benchmark in February 2025.
One of the things we measured is the accuracy of citations. Each answer was given a score out of three on citations and an answer containing a fictional citation was automatically given a score of 0. The table below provides an overview of how often the LLMs we tested were awarded a score of zero on this basis.
These numbers are broadly consistent with the Ayinde judgment below where, in one of the situations considered by the Divisional Court, the claimant listed 45 cases, of which 18 did not exist.
Having said that, the most recent LLMs we tested (Gemini and o1) are improving and only included fictitious cases in 9% of the questions we asked. The accuracy of these tools should improve further over time, and newer tools offer research capabilities to help check the citations provided. However, there is clearly still a problem.
The Divisional Court’s judgment in Ayinde, R v Qatar National Bank QPSC & Anor [2025] EWHC 1383 is a textbook example of why this really matters in practice. The court indicated that including fictitious citations in pleadings is an extremely serious matter that might well lead to:
The Divisional Court suggests that, at a minimum, the inclusion of fake citations is likely to result in a referral to a regulator and a wasted costs award. While a public admonishment could be issued instead, given the risks to the administration of justice, admonishment alone is unlikely to be a sufficient response.
Perhaps more significant is the pointed remarks to those managing law firms or chambers. The court states “practical and effective measures must now be taken by those within the legal profession with individual leadership responsibilities (such as heads of chambers and managing partners) and by those with the responsibility for regulating the provision of legal services…For the future… the profession can expect the court to inquire whether those leadership responsibilities have been fulfilled”.
The court’s strong reaction is partly a product of the extremely poor handling of one of the cases before it, which related to judicial proceedings against the London Borough of Haringey. The grounds for judicial review were settled by counsel (Ms F) and contained no less than five fictitious cases.
When Haringey’s solicitors identified the issue, the correct response would have been to immediately investigate and provide a full and comprehensive apology. Instead, Ms F drafted an extraordinary response variously stating that the claimants did “not see the point you are making by correlating any errors in citations to the issues addressed in the request for judicial review”. The response went on: “We hope that you are not raising these errors as technicalities to avoid undertaking really serious legal research …It appears to us improper to barter our client's legal position for cosmetic errors”.
When the matter came before Ritchie J ([2025] EWHC 1040 (Admin)), Ms F denied using artificial intelligence tools claiming that “she kept a box of copies of cases, and she kept a paper and digital list of cases with their ratios. She said that she had "dragged and dropped" the reference to [the fictitious cases] from that list into the grounds for judicial review.”
When the matter came before the Divisional Court, Ms F continued to deny using artificial intelligence tools, though claimed she “"may also have carried out searches on Google or Safari" and that she may have taken account of artificial intelligence generated summaries of the results”.
The Divisional Court was wholly unimpressed, concluding that Ms F either deliberately included fake citations in her written work or used generative artificial intelligence tools and lied about it. Both possibilities met the threshold for initiating contempt proceedings.
In the end the court decided not to initiate contempt proceedings, mainly because Ms F was extremely junior. It instead referred Ms F to the Bar Standard Board and referred the instructing solicitor to the Solicitor Regulation Authority (on top of Richie J’s earlier award of wasted costs).
However, the Divisional Court ends with a severe warning. Their “decision not to initiate contempt proceedings in respect of Ms [F] is not a precedent. Lawyers who do not comply with their professional obligations in this respect risk severe sanction”.
There are a number of immediate practical lessons: