Lawyer uses ChatGPT in court and now ‘greatly regrets’ it

The court used the word “bogus” to describe the quotes and citations the New York attorney provided, which were sourced from ChatGPT.

A New York attorney has been blasted for using ChatGPT for legal research as part of a lawsuit against a Columbian airline.

Steven Schwartz, an attorney with New York law firm Levidow, Levidow & Oberman, was engaged by Robert Mata to pursue an injury claim against Avianca Airlines.

Mata claims he sustained the injury from a serving cart during his flight with the airline in 2019, according to a May 28 report from CNN Business.

However, after a judge noticed inconsistencies and factual errors in the case documentation, Schwartz has now admitted to using ChatGPT for his legal research, according to a May 24 sworn affidavit.

He claims that this was his first time using ChatGPT for legal research and “was unaware of the possibility that its content could be false.”

In an April 5 court filing, the judge presiding over the case stated:

“Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations.”

The judge further claimed that certain cases referenced in the submissions did not actually exist, and there was an instance where a docket number on a filing was mixed up with another court filing. 

Extract of Steven Schwartz’s sworn affidavit on May 24. Source: courtlistener.com

Schwartz said he also regrets having trusted the artificial chatbot without conducting his own due diligence. The affidavit noted:

“Greatly regrets having utilized generative artificial intelligence to supplement the legal research performed herein and will never do so in the future without absolute verification of its authenticity.”

Related: AI meets blockchain: Revolutionizing smart contracts and cryptocurrency

In recent times there has been an ongoing debate regarding the extent to which ChatGPT can be integrated into workforces.

However, reports indicate that the intelligence levels of ChatGPT is rapidly advancing.

However, developers are skeptical over whether it does in fact have the potential to replace humans completely. 

Syed Ghazanfer, a blockchain developer, said while he is in favor of ChatGPT, he is doubtful that it has the communication skills to completely replace human workers.

“For it to replace you, you have to communicate requirements which are not possible in native English. That’s why we invented programming languages” he said.

Magazine: ‘Moral responsibility’: Can blockchain really improve trust in AI?

Source Link

Share with your friends!

Products You May Like

Leave a Reply

Your email address will not be published. Required fields are marked *