Home » Blog » AI Evidence in Courts: Should Algorithms Be Treated as Expert Witnesses

AI Evidence in Courts: Should Algorithms Be Treated as Expert Witnesses

Authored By: Khadija Bilal

Pakistan College of Law

ABSTRACT 

The increasing reliance on artificial intelligence (AI) and algorithmic tools in judicial processes  has raised complex questions regarding the admissibility, reliability, and accountability of such  evidence. Courts across jurisdictions are beginning to rely on algorithmic risk assessments,  predictive tools, and data-driven systems, particularly in criminal justice contexts such as bail,  sentencing, and forensic analysis. This article critically examines whether AI-generated outputs  should be treated as expert evidence within existing legal frameworks. It argues that while  algorithmic tools may enhance efficiency and consistency, their opaque functioning,  susceptibility to bias, and limited explainability pose serious challenges to fair trial rights,  judicial transparency, and accountability. The article concludes that AI should function strictly  as a decision-support mechanism and must not be equated with human expert witnesses unless  robust legal safeguards are established. 

  1. INTRODUCTION 

Technology has always had an impact on how justice is served. Courts are constantly adapting  to new types of evidence, like fingerprinting, DNA analysis, CCTV, and digital forgery. The  current “AI turn” in legal systems is more disruptive than earlier changes because it doesn’t just  collect information; it also processes it, draws conclusions, and sometimes makes suggestions  that change the outcomes.1Increasingly, tools with names like “risk assessment,” “predictive  analytics,” or “decision support” are being used in the criminal justice system to help with bail  decisions, sentencing, and deciding where to put investigative resources.2 

This trend raises a challenging legal issue: should algorithmic outputs to be regarded as expert  evidence? Expert testimony is allowed because it helps the court understand things that most  people don’t know about. But experts have to follow certain rules: they have to use a recognised  method, explain their reasoning, and be able to be cross-examined.3 Different AI systems work in different ways. They might use confidential code and training data, and even the people who  set them up might not be able to understand how they work.4 

It’s not just that “technology is new” that worries people; it’s also that algorithmic evidence can  undermine the justice system’s core values, like the right of parties to test evidence, understand  why a decision is made, and hold decision-makers accountable. These promises are especially  important in criminal cases, where the results can mean losing your freedom and being shamed.  The right to a fair trial, which is protected by due process principles in legal systems and Article  6 of the European Convention on Human Rights (ECHR) in Europe, means that people should  have a real chance to question the evidence and reasoning used by courts.5 

This article argues that AI-generated outputs should not be treated as expert witnesses by  default. The better view is that algorithmic tools can help people make decisions, but they  should only be allowed to be used in very specific ways. When AI evidence is used, courts  need to put in place protections that promote disclosure, explainability, bias assessment, and  contestability. If they don’t, algorithmic evidence could become an “unanswerable expert,”  which would hurt both fairness and public trust in the legal system. 

  1. RESEARCH METHODOLOGY 

This article employs a doctrinal and analytical approach. It is based on: (i) the principles that  govern expert evidence in common law, (ii) case law that shows the risks and limits of technical  evidence that is too persuasive, and (iii) fair trial standards based on human rights that stress  openness and the right to challenge evidence. Comparative insight is employed to demonstrate  the practical judicial interaction with algorithmic tools, notably through the case of State v.  Loomis in the United States.6 The goal is not to say that all AI evidence is not allowed, but to  make clear what conditions need to be met before algorithmic outputs can be safely used in  court. 

  1. EXPERT EVIDENCE AND ITS LEGAL LIMITS 

Expert evidence is an exception to the general rule that witnesses can only talk about things  they know and not make judgements about them. Experts are allowed to help the court with specific issues. But because testimony from experts can be very convincing, courts have always  required safeguards to make sure it doesn’t change the way cases are decided. 

The English Court of Appeal said in R v. Turner that experts should not “usurp the function”  of the judge or jury. If the jury can make up its own mind, expert opinion may not be needed  or appropriate.7 The point is constitutional in a small “c” sense: decision making authority rests  with the court, not with external specialists. Expert evidence is justified only when it genuinely  assists and remains open to scrutiny. 

R v Bonython, a well-known Australian case that is often cited for its clarity, also lists two  main questions about admissibility: (1) is the subject matter such that a lay tribunal needs help,  and (2) does the witness have enough expertise based on a recognised body of knowledge?8 The second limb is important because it connects admissibility to reliability. Expertise isn’t just  confidence or credentials; it’s disciplined knowledge that can be evaluated. 

Cross-examination is an important part of the process that protects people. Experts can tell you  why they think what they do, what they think, what they think is wrong, and how to deal with  criticism. Cross-examination therefore helps courts distinguish solid expertise from persuasive  guesswork. It also promotes fairness by giving the other side a fair chance to challenge the  “special” evidence that was used. 

The legal model of expert evidence thus assumes three elements: (i) a rational methodology;  (ii) transparency regarding the derivation of conclusions; and (iii) the capacity for contestation  via cross-examination and rebuttal. When the “expert” is an algorithm, these assumptions start  to break down. 

  1. WHAT MAKES AI EVIDENCE DIFFERENT

AI systems that are often used in courts and other places of justice are not conscious thinkers;  they are tools for finding patterns in data. Their outputs are based on the data they were trained  on, the way the model is set up, and the choices made about the parameters. A tool may not  work well in certain situations, even if it works well on average. This is especially true when  the training data does not fairly represent the population.  

  • The Black Box Problem 

Many contemporary machine learning systems, particularly those employing intricate models,  generate outputs that are challenging to elucidate in human terms. The critique of the “black  box” is not merely philosophical; it is substantiated by evidence. A party cannot effectively  contest an algorithmic conclusion if they do not comprehend its foundation.9 The court also  can’t give the evidence any weight unless it knows what its limits are. 

In traditional evidence law, reasoning, methodological scrutiny, and adversarial challenge are  used to test reliability. Algorithmic opacity messes up these steps. AI evidence can look like  an “answer” without a visible “working,” which is the opposite of what courts usually want  from experts. 

  • Bias and Historical Data 

AI tools that learn from past data can copy the biases that are already in that data. If police have  historically focused on certain neighbourhoods, arrest data may show bias in enforcement  rather than actual crime rates. When risk tools are trained on this kind of data, they might call  people from communities that are heavily policed “higher risk,” which makes the inequality  worse. 10 

Saying “the tool is neutral” doesn’t help with bias issues. The claim of neutrality ignores the  fact that design choices decide what counts as risk, which variables to include, and which  outcomes to optimise. Those choices are loaded with meaning, but they are often hidden behind  technical language. 

  • Automation Bias and Deference 

Even when a judge is not legally bound by an algorithm, algorithmic outputs can carry  psychological authority. Judges may defer to a numeric score or recommendation because it  appears objective and scientific. This “automation bias” can shift the real centre of gravity of  a decision away from judicial reasoning, even while formal discretion remains.11 The danger is not that judges stop deciding, but that their decisions become anchored by tools they cannot  fully interrogate. 

  1. ADMISSIBILITY, RELIABILITY, AND THE LESSONS OF TECHNICAL EVIDENCE 

Courts already know that technical proof can be very convincing in a bad way. The Court of  Appeal in R v. Doheny stressed the need for careful directions and caution when dealing with  scientific proof (in this case, DNA evidence) because juries may give it too much weight.12 This warning is clear in algorithmic terms: “technical” doesn’t mean “infallible,” and courts  shouldn’t treat complicated tools as if they are always right. 

If algorithmic outputs are introduced as expert evidence, the court must ask: (i) what is the  methodology; (ii) can it be scrutinised; (iii) what is the error rate; (iv) how was the tool  validated; (v) does it perform differently across groups; and (vi) can the opposing party  test it? These are not optional quest Yet, AI companies often say that they own the code and data that they use. If commercial  confidentiality prevents disclosure, the adversarial system’s capacity to assess reliability is  diminished. This creates a structural unfairness because one party (usually the state) has access  to the tool’s authority, but the other party does not have access to its foundations. 

This is why it’s dangerous to treat AI outputs as “expert evidence.” It can bring unclear  reasoning into court under the guise of expertise, without the protections that make expertise  legitimate. 

  1. FAIR TRIAL AND DUE PROCESS 

Fair trial guarantees demand that individuals have a meaningful opportunity to challenge  evidence relied upon in determining their rights. Article 6 ECHR, for example, reflects  principles of procedural fairness and equality of arms.13 Even outside ECHR systems, similar  values appear through constitutional due process and common law fairness. 

A system that relies on algorithmic outputs that cannot be challenged risks violating these  guarantees. If a defendant is told, “the algorithm says you are high risk,” but cannot see the  basis, the model, or the assumptions, the ability to challenge becomes superficial The European Court of Human Rights has said many times how important it is to have  procedural protections and make decisions based on reasons. Osman v. United Kingdom is not  an “AI case,” but it shows how important fairness norms are under Article 6.14 Courts have to  be competent to explain their decisions in a way that people can understand, check, and appeal.  AI systems that give outputs without clear reasons go against that goal. 

There is also a problem with legitimacy. Courts get their power not just from being right, but  also from being able to explain their decisions in a way that makes sense to the public. People  may not be able to trust court decisions as much when they are made by systems that aren’t  visible. This is especially troubling in the criminal justice system, where people need to believe  that outcomes are reached through fair processes, not hidden machinery. 

  1. COMPARATIVE PERSPECTIVE: STATE V LOOMIS 

State v. Loomis is a well-known case that shows how algorithmic decision-making works. In  this case, a sentencing court used a proprietary risk assessment tool.15 The Wisconsin Supreme  Court said it was okay to use, but it was worried about transparency and said it shouldn’t be the  only thing used to decide sentences. 

Loomis shows the main problem: algorithmic tools can help keep things consistent, but their  lack of transparency can lead to unfairness. The case is also important because it shows that  judges know that “admitting” algorithmic outputs is not the end of the analysis. The court must  also say how the outputs can be used, what warnings go with them, and what limits they have. 

But warnings alone aren’t enough if the defendant still can’t challenge the tool in a meaningful  way. A warning that “this tool may be biassed” does not help a defendant challenge the details  of how the score was made. The case thus corroborates the thesis of this article: algorithmic  tools may be employed, but they should not be treated as conventional experts unless  genuine contestability exists. 

  1. SHOULD ALGORITHMS BE TREATED AS EXPERT WITNESSES

Treating AI as an “expert witness” is conceptually tempting because AI seems to do what  experts do: analyse complex information and generate conclusions. But legally, the analogy  breaks down. 

First, an “expert witness” is responsible. A human expert can be criticised, discredited, and  punished in their field. An algorithm cannot be morally or professionally accountable.  Developers, vendors, and users are now responsible, but evidential law isn’t meant to spread  responsibility like that. 

Second, experts can be questioned in a cross-examination. Algorithms are not able to answer  questions. A human representative can explain the system at best, but that doesn’t mean they  can test how the system thinks, especially if the explanations are incomplete or come after the  fact. 

Third, experts are expected to reason. Many AI tools do not “reason” in a way courts can  recognise. They classify or predict based on correlations that may not be causally meaningful. 

For these reasons, elevating AI to expert status risks creating an “unchallengeable” authority.  The safer legal stance is: AI outputs should be treated as technical material that must be  supported by a human expert who can explain validation, limitations, bias testing, and  relevance, and whose testimony remains contestable. Even then, courts should be cautious  about how much weight is assigned. 

  1. WAY FORWARD: SAFEGUARDS AND REFORM 

If courts are to use AI evidence responsibly, the law must develop safeguards that translate  fairness principles into operational rules. 

  • Transparency and Disclosure 

Courts should mandate the disclosure of training data characteristics (at a minimum, in  summary form), validation studies, established error rates, and the model’s intended application.  When proprietary claims hinder disclosure, courts should evaluate whether fairness  necessitates exclusion or diminished weight. 

  • Explainability Standards

When an output impacts liberty or significant rights, courts ought to require a level of  explainability adequate for substantive challenge. Be very careful with a “risk score” that  doesn’t come with an explanation.  

  • Bias Audits and Equality 

Before use, there should be independent bias audits. Courts should also look at whether an  algorithm works differently for different groups that are protected and whether using it makes  structural inequality worse. 

  • Judicial Training 

Judges don’t need to be engineers, but they should know how to ask simple questions about  bias, reliability, error rates, and validation. Without this, courts might rely on technical  authority without fully understanding it. 

  • Human Oversight and Limits on Use 

AI should always be helpful. Courts should not let algorithmic outputs decide things. Instead,  they should require independent judicial reasoning and give clear reasons when they do use  algorithmic outputs. 

  1. CONCLUSION 

AI tools can support courts, but they also risk importing opaque, biased, and unchallengeable  reasoning into legal decisions. Traditional expert evidence is tolerated because it is transparent,  contestable, and accountable. Many algorithmic tools fail those conditions. Accordingly, AI generated outputs should not be treated as expert witnesses by default. Instead, courts must  insist on disclosure, explainability, bias auditing, and the preservation of meaningful challenge.  The guiding principle should be simple: technology must serve justice, not replace the  conditions that make justice legitimate. 

BIBLIOGRAPHY 

Cases 

  • Osman v United Kingdom (1998) 29 EHRR 245 
  • R v Doheny [1997] 1 Cr App R 369
  • R v Turner [1975] QB 834 
  • R v Bonython (1984) 38 SASR 45 
  • State v Loomis 881 NW 2d 749 (Wis 2016) 

Legislation and Materials 

  • European Convention on Human Rights (ECHR) art 6. 

Books and Reports 

  • Richard Susskind, Tomorrow’s Lawyers: An Introduction to Your Future (2nd edn, OUP  2017). 

Journal Articles 

  • Danielle Keats Citron, ‘Technological Due Process’ (2008) 85 Washington University Law  Review 1249. 
  • Frank Pasquale, ‘A Rule of Persons, Not Machines’ (2019) 87 George Washington Law  Review 1.

1Richard Susskind, Tomorrow’s Lawyers: An Introduction to Your Future (2nd edn, OUP 2017)

2 State v Loomis 881 NW 2d 749 (Wis 2016) 

3 R v Bonython (1984) 38 SASR 45

4 Frank Pasquale, ‘A Rule of Persons, Not Machines’ (2019) 87 George Washington Law Review 1

5 European Convention on Human Rights (ECHR) art 6 

6 State v Loomis 881 NW 2d 749 (Wis 2016)

7 R v Turner [1975] QB 834 

8 R v Bonython (1984) 38 SASR 45

9 Danielle Keats Citron, ‘Technological Due Process’ (2008) 85 Washington University Law Review 1249

10 Citron (n 9) 

11 Pasquale (n 4)

12 R v Doheny [1997] 1 Cr App R 369 

13 ECHR art 6

14 Osman v United Kingdom (1998) 29 EHRR 245 

15 State v Loomis 881 NW 2d 749 (Wis 2016)

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top