We've seen fake, explicit pictures of Celebrities, created by artificial intelligence (AI). AI has also played a task in the event of music, driverless racing cars, and the spread of misinformation, amongst other things.
It is subsequently hardly surprising that AI can be having a robust impact on our legal systems.
It is common knowledge that courts must resolve disputes based on the law that lawyers undergo the court as a part of a client's case. Therefore, it is amazingly worrying that fake laws invented by AI are getting used in litigation.
This not only raises questions of legality and ethics, but in addition threatens to undermine faith and trust in global legal systems.
How do fake laws come about?
There is little doubt that generative AI is a strong tool with transformative potential for society, including many points of the legal system. But its use brings with it responsibility and risks.
Lawyers are trained within the careful application of experience and experience and usually don’t take major risks. Some careless lawyers (and represented himself Litigants) were caught by artificial intelligence.
AI models are trained on huge data sets. When prompted by a user, they’ll create recent content (each text and audiovisual content).
Although content generated this fashion can look very compelling, it might even be inaccurate. This is the results of the AI model attempting to “fill within the gaps” when its training data is insufficient or flawed, and is often known as “hallucination“.
In some contexts, generative AI hallucination shouldn’t be an issue. In fact, it might be viewed for instance of creativity.
But when AI hallucinates or creates inaccurate content that’s then utilized in legal proceedings, that could be a problem – especially when combined with time constraints for lawyers and a scarcity of access to legal services for many individuals.
This powerful combination can result in negligence and shortcuts in legal research and document preparation, potentially resulting in reputational problems for the legal career and a scarcity of public trust within the administration of justice.
It's already happening
The best-known “fake case” of generative AI is the US case from 2023 Mata vs Avianca, wherein lawyers submitted a transient containing fake excerpts and case citations to a New York court. The order was researched using ChatGPT.
Because the lawyers didn’t know that ChatGPT could hallucinate, they did not confirm whether the cases actually existed. The consequences were devastating. After the error was discovered, the court dismissed her client's case, sanctioned the lawyers for bad faith, fined them and their firm, and subjected their actions to public scrutiny.
Despite negative publicity, more fake case studies proceed to emerge. Michael Cohen, Donald Trump's former lawyer, presented his own legal cases generated by Google Bard, one other generative AI chatbot. He believed they were real (they weren't) and that his lawyer would fact check them (he didn't). His lawyer including the cases in a transient filed in U.S. federal court.
In recent cases, fake cases have also emerged Canada And the UK.
If this trend goes unchecked, how can we be certain that the careless use of generative AI doesn’t undermine public trust within the legal system? Consistent failure by lawyers to exercise due care when using these tools can mislead and overburden the courts, harm clients' interests and usually undermine the rule of law.
What is being done about it?
Around the world, regulators and courts have responded in alternative ways.
Several US states and courts have issued guidelines, opinions or orders on using generative AI, starting from responsible adoption to an outright ban.
Law societies within the United Kingdom and British Columbia and the New Zealand courts have also developed guidelines.
In Australia the NSW Bar Association has one Generative AI Guide for lawyers. The Bar Association of NSW and that Law Institute of Victoria have published articles on responsible behavior in accordance with the Code of Conduct for Lawyers.
Many lawyers and judges, like the general public, could have some understanding of generative AI and recognize each its limitations and its advantages. But there are others who is probably not so aware of it. Instructions undoubtedly help.
However, a prescriptive approach is required. Lawyers using generative AI tools cannot consider this as an alternative choice to exercising their very own judgment and care and must confirm the accuracy and reliability of the data received.
In Australia, courts should adopt practice guidance or rules setting out expectations when using generative AI in litigation. Court rules can even provide guidance to self-represented litigants and would convey to the general public that our courts are aware of the problem and are addressing it.
The legal career could also adopt formal guidelines to advertise the responsible use of AI by lawyers. At the very least, technology competency should change into a requirement for continuing legal education for lawyers in Australia.
Establishing clear requirements for the responsible and ethical use of generative AI by lawyers in Australia will encourage appropriate adoption and increase public confidence in our lawyers, our courts and the general administration of justice on this country.