WeekendTimes.com.au



Men's Weekly

.

AI is creating fake legal cases and making its way into real courtrooms, with disastrous results

  • Written by Michael Legg, Professor of Law, UNSW Sydney
AI is creating fake legal cases and making its way into real courtrooms, with disastrous results

We’ve seen deepfake, explicit images of celebrities[1], created by artificial intelligence (AI). AI has also played a hand in creating music[2], driverless race cars[3] and spreading misinformation[4], among other things.

It’s hardly surprising, then, that AI also has a strong impact on our legal systems.

It’s well known that courts must decide disputes based on the law, which is presented by lawyers to the court as part of a client’s case. It’s therefore highly concerning that fake law, invented by AI, is being used in legal disputes.

Not only does this pose issues of legality and ethics, it also threatens to undermine faith and trust in global legal systems.

Read more: Lawyers are rapidly embracing AI: here's how to avoid an ethical disaster[5]

How do fake laws come about?

There is little doubt that generative AI is a powerful tool with transformative potential for society, including many aspects of the legal system. But its use comes with responsibilities and risks.

Lawyers are trained to carefully apply professional knowledge and experience, and are generally not big risk-takers. However, some unwary lawyers (and self-represented[6] litigants) have been caught out by artificial intelligence.

ChatGPT on a smartphone screen in front of the same website on a laptop screen
Generative AI tools, like ChatGPT, can provide incorrect information. Shutterstock[7]

AI models are trained on massive data sets. When prompted by a user, they can create new content (both text and audiovisual).

Although content generated this way can look very convincing, it can also be inaccurate. This is the result of the AI model attempting to “fill in the gaps” when its training data is inadequate or flawed, and is commonly referred to as “hallucination[8]”.

In some contexts, generative AI hallucination is not a problem. Indeed, it can be seen as an example of creativity.

But if AI hallucinated or created inaccurate content that is then used in legal processes, that’s a problem – particularly when combined with time pressures on lawyers and a lack of access to legal services for many.

This potent combination can result in carelessness and shortcuts in legal research and document preparation, potentially creating reputational issues for the legal profession and a lack of public trust in the administration of justice.

It’s happening already

The best known generative AI “fake case” is the 2023 US case Mata v Avianca[9], in which lawyers submitted a brief containing fake extracts and case citations to a New York court. The brief was researched using ChatGPT.

The lawyers, unaware that ChatGPT can hallucinate, failed to check that the cases actually existed. The consequences were disastrous. Once the error was uncovered, the court dismissed their client’s case, sanctioned the lawyers for acting in bad faith, fined them and their firm, and exposed their actions to public scrutiny.

Read more: AI is everywhere – including countless applications you've likely never heard of[10]

Despite adverse publicity, other fake case examples continue to surface. Michael Cohen, Donald Trump’s former lawyer, gave his own lawyer cases generated by Google Bard, another generative AI chatbot. He believed they were real (they were not) and that his lawyer would fact check them (he did not). His lawyer included the cases[11] in a brief filed with the US Federal Court.

Fake cases have also surfaced in recent matters in Canada[12] and the United Kingdom[13].

If this trend goes unchecked, how can we ensure that the careless use of generative AI does not undermine the public’s trust in the legal system? Consistent failures by lawyers to exercise due care when using these tools has the potential to mislead and congest the courts, harm clients’ interests, and generally undermine the rule of law.

A man in a suit leaves a courtroom Michael Cohen’s lawyer was caught up in a court case involving fake AI case law. Sarah Yenesel/EPA

What’s being done about it?

Around the world, legal regulators and courts have responded in various ways.

Several US state bars and courts have issued guidance, opinions or orders on generative AI use, ranging from responsible adoption to an outright ban.

Law societies in the UK and British Columbia, and the courts of New Zealand, have also developed guidelines.

In Australia, the NSW Bar Association has a generative AI guide[14] for barristers. The Law Society of NSW[15] and the Law Institute of Victoria[16] have released articles on responsible use in line with solicitors’ conduct rules.

Many lawyers and judges, like the public, will have some understanding of generative AI and can recognise both its limits and benefits. But there are others who may not be as aware. Guidance undoubtedly helps.

But a mandatory approach is needed. Lawyers who use generative AI tools cannot treat it as a substitute for exercising their own judgement and diligence, and must check the accuracy and reliability of the information they receive.

Read more: Do you trust AI to write the news? It already is – and not without issues[17]

In Australia, courts should adopt practice notes or rules that set out expectations when generative AI is used in litigation. Court rules can also guide self-represented litigants, and would communicate to the public that our courts are aware of the problem and are addressing it.

The legal profession could also adopt formal guidance to promote the responsible use of AI by lawyers. At the very least, technology competence should become a requirement of lawyers’ continuing legal education in Australia.

Setting clear requirements for the responsible and ethical use of generative AI by lawyers in Australia will encourage appropriate adoption and shore up public confidence in our lawyers, our courts, and the overall administration of justice in this country.

References

  1. ^ celebrities (www.nytimes.com)
  2. ^ creating music (theconversation.com)
  3. ^ driverless race cars (theconversation.com)
  4. ^ misinformation (theconversation.com)
  5. ^ Lawyers are rapidly embracing AI: here's how to avoid an ethical disaster (theconversation.com)
  6. ^ self-represented (reason.com)
  7. ^ Shutterstock (www.shutterstock.com)
  8. ^ hallucination (www.csiro.au)
  9. ^ Mata v Avianca (law.justia.com)
  10. ^ AI is everywhere – including countless applications you've likely never heard of (theconversation.com)
  11. ^ included the cases (www.reuters.com)
  12. ^ Canada (www.cbc.ca)
  13. ^ the United Kingdom (www.legalfutures.co.uk)
  14. ^ generative AI guide (inbrief.nswbar.asn.au)
  15. ^ Law Society of NSW (lsj.com.au)
  16. ^ Law Institute of Victoria (www.liv.asn.au)
  17. ^ Do you trust AI to write the news? It already is – and not without issues (theconversation.com)

Authors: Michael Legg, Professor of Law, UNSW Sydney

Read more https://theconversation.com/ai-is-creating-fake-legal-cases-and-making-its-way-into-real-courtrooms-with-disastrous-results-225080

The Weekend Times Magazine

What’s the Difference Between a Caravan & A Motorhome?

Australians love the freedom of the open road, and choosing the right setup can make travelling safer and far more enjoyable. With both caravans and motorhomes growing in popularity and...

Could You Furnish and Equip a Whole House with IKEA Products?

IKEA has long been known as the global giant of affordable, ready-to-assemble furniture. From minimalist Scandinavian sofas to clever kitchen gadgets, the Swedish retailer has built its reputation on making...

The Future of the Sales Handoff: From AI SDR to Human Closer

Artificial intelligence is altering the sales process at lightning speed. For many firms, AI is their Sales Development Representative. This class of SDRs completes the initial stage of the sales...

Property app Instarent

Property self-management soars during COVID lockdown The innovative PropTech app, Instarent, has seen exponential growth during the COVID -19 lockdown, reporting a 400 per cent increase in users during...

Smart Lock: The Future of Home Security and Convenience

A smart lock has revolutionized the way people think about home security. Moving beyond traditional keys and mechanical locks, smart locks bring technology and convenience together, offering homeowners and businesses a new...

How To Install PVC Shutters At Home

Homes are the best place for relaxing and refreshing after the hectic and tiring daily life routine. The home should be decorated so that it gives people a warm welcome...

Strong Australia panel interview with Kieran Gilbert

Kieran Gilbert, chief news anchor Sky News: The Business Council of Australia continued its Strong Australia series today. This time the spotlight on the city of Wagga. How are regional cities...

Brisbane Compensation Lawyers for Work Injury Damages

Suffering a work-related injury can turn your life upside down. The physical pain, emotional stress, and financial uncertainty can feel overwhelming. If you’re in Brisbane and facing this situation, finding...

Baking Tools and Equipment Your Bakery Needs

It can be hard to resist the smell of fresh bread or devouring a freshly baked cake. Fortunately, some people have a knack for kneading dough and baking up a...