However, generative AI technologies could be beneficial for legal work. Lawyers do generate lots of text. In fact, at our law firm, we are using several solutions that assist with legal work and business legal operations. But we are careful and train our lawyers to use these tools properly.

The ChatGPT cases have sparked debates about the ethical and professional implications of using generative AI tools in legal practice and court. Some courts have now added requirements that any use of AI have a certification by the lawyer that confirms they have verified all the cases cited. The Chief Justice of the BC Supreme Court issued a directive telling judges not to use AI. State bars and law societies have said that lawyers should be more careful and responsible when incorporating technology into their work. Recently, the State bar of Florida approved an Ethics Advisory Opinion 24-1, which says that lawyers may ethically use generative AI if they can guarantee compliance with the lawyer’s ethical obligations, including duty to client confidentiality. I agree with many commentators that generative AI tools could be helpful and innovative if used correctly and with adequate training.

I have met many lawyers who are wary of generative AI solutions because of cases like Schwartz’s and the problematic nature of hallucinations. However, advancements in AI technology are quickly emerging to address this issue. Retrieval-augmented generation (RAG) is one such solution. RAG is a technique that enhances the output of large language models (LLMs), like GPT, by retrieving relevant information from external sources. RAG aims to overcome the limitation of hallucinations by providing LLMs with additional context and facts that improve their responses’ accuracy, reliability, and trustworthiness. Think of RAG as an assistant for LLMs, retrieving relevant case law and precedents. Systems that employ RAG will cite sources that users can easily click on and confirm that the text generated is correct, reducing inaccuracies and improving trust in generative AI technologies.

There are already several legal tech solutions emerging that incorporate the use of RAGs and do not have the problem of hallucinations. This technology will likely significantly impact legal research and document drafting, which can be very time-consuming tasks. Instead, imagine drafting a legal research memo that pulls your evidence and caselaw with appropriate citations.

Lexis+ and Paxton.AI are examples of legal tech solutions incorporating RAG technology. As AI technology evolves quickly, it may soon render any attempts by courts and law societies to limit AI use irrelevant. It appears that technology will solve its problems.



Source link

author-sign