2025年12月12日
AI disputes in action – 2 / 1 观点
Many aspects of AI are now commonplace in UK Courts, from computer-assisted disclosure to assisting judges in the preparation of their judgments. The rapid development of AI poses a new challenge, however: how will the courts deal with the inevitable volume of AI-generated material that will be placed before them and how will they distinguish between genuine and synthetic evidence?
As noted in the recently updated Guidance for Judicial Office Holders, the judiciary is aware that "AI tools are now being used to produce fake material, including text, images and video" but that "Courts and tribunals have always had to handle forgeries […] involving varying levels of sophistication". On any view, however, courts have not faced forgeries of the nature, sophistication and scale now potentially posed. The rapid advancement in AI tools has arguably transformed the scope for forgery from a relatively rare and detectable anomaly into a widely available, existential threat to evidential integrity.
Some of the primary challenges posed by AI-generated evidence are that:
The courts' existing approach to allegations of metadata interference may be instructive for how allegations of inauthentic AI-generated evidence will be addressed by the judiciary. Both issues share common challenges around authenticity and provenance, technical complexity often requiring forensic expert evidence, burden of proof considerations (a party calling such evidence into question would likely be required to provide credible evidence to support such allegations) and the amount of weight that may be placed on evidence where integrity has been questioned (if even it is admitted).
Courts are likely to adopt a cautious approach to suggestions of AI forgeries, heavily informed by expert evidence, just as they have done when addressing metadata tampering allegations. In reality, however, even if the approach to assessing metadata integrity will provide a basic framework for assessing the authenticity of AI-generated material, the development and implementation of a specific framework for AI evidence is likely to be necessary, not least because AI evidence will leave a far less concrete digital trace than metadata.
AI-generated material may be legitimately provided to the court (AI-produced summaries of voluminous data that might otherwise require expensive expert analysis, for example). Such outputs will undoubtedly not only aid the judiciary but also potentially assist in democratising litigation, allowing claimants to bring claims and strategically litigate against bigger players with deeper pockets when they might otherwise be restricted in bringing a claim by the costs of expert input or processing vast quantities of highly technical information.
The output from AI in this way would not, however, amount to evidence itself – instead, in this context, AI would be an evidence and data-processing tool and a human would need to take responsibility for its interpretation. In this way, courts might theoretically treat the output in a similar way to how they have now widely accepted technology-assisted review (TAR) and predictive coding in disclosure exercises (using machine learning algorithms), through for example:
In theory, therefore, proportionality, rather than burdensome technical explanations, might sit at the heart of the courts' approach to AI evidence-processing. The reality is likely to look very different, however – in particular, being able to demonstrate human oversight, training and decision-making may be impossible when it comes to using AI as a processing tool. If the courts do in due course expect to interrogate the AI's training, accuracy, and algorithmic biases, it will undoubtedly add complexity to proceedings that AI may have been intended to simplify. Moreover, and as the Guidance for Judicial Office Holders notes, malicious actors can embed invisible 'white text' that is imperceptible to humans but detectable by AI to manipulate outputs through hidden instructions. This means that what may appear at face value to be neutral AI-generated outputs may instead have been covertly skewed, creating a new form of fraud that may be extremely difficult to detect.
If, on the other hand, an AI system generated opinions from data, it would effectively be akin to an expert rather than merely a processing tool (where a human would ultimately remain responsible for its interpretations) and its output (eg predictions, diagnoses, technical judgements or causation analysis) would resemble expert opinion evidence.
It is difficult to see how a court would reconcile the requirements of CPR 35 (which covers expert evidence) in such instance – who, for example, is the expert? Is it the AI system itself, the developer, the human who 'ran' it, or someone else? If it was the system, how would that 'expert' satisfy the court that it was qualified, objective and impartial? Absent a human explanation of whether methods were reliable and why conclusions were reached, it is hard to envisage AI "evidence" being admitted or given much weight.
The courts do of course already deal with machine-generated expertise (in some highly technical and scientific areas), but although the underlying expertise emanates from a system, the output only becomes admissible as 'expert evidence' when a human expert understands it, assumes responsibility for it, and can be cross-examined in relation to it. Without this, the AI output is little more than hearsay from a machine. For now, we seem very far indeed from AI systems becoming experts whose opinions are admissible in their own right.
Left with the more immediate possibility and challenge of AI-generated fake evidence, we foresee that the following potential impacts may materialise in the near future:
As cited in the Law Commission's 2025 Discussion Paper, even where the technical details of AI systems are available, AI systems are challenging even for AI experts to explain as they are based on mathematical models of enormous scale and, as leading AI developer Anthropic has stated, "a surprising fact about modern large language models is that nobody really knows how they work internally". This fundamental opacity suggests that the key challenge that judges, lawyers and parties will grapple with, and which will shape the future of disputes, is how courts and tribunals, predicated on establishing facts and evidence, will identify or even accommodate evidence from processes that are, by their creators' own admission, inherently inexplicable.