ChatGPT cited “bogus” cases for a New York federal court filing

Celebrity Gig


The Thurgood Marshall courthouse is pictured in Manhattan in New York, October 15, 2021.

Brendan McDermid | Reuters

Roberto Mata’s lawsuit against Avianca Airlines wasn’t so different from many other personal-injury suits filed in New York federal court. Mata and his attorney, Peter LoDuca, alleged that Avianca caused Mata personal injuries when he was “struck by a metal serving cart” on board a 2019 flight bound for New York.

Avianca moved to dismiss the case. Mata’s lawyers predictably opposed the motion and cited a variety of legal decisions, as is typical in courtroom spats. Then everything fell apart.

Avianca’s attorneys told the court that it couldn’t find numerous legal cases that LoDuca had cited in his response. Federal Judge P. Kevin Castel demanded that LoDuca provide copies of nine judicial decisions that were apparently used.

READ ALSO:  Expose ChatGPT Resumes and Uncover Real Talent Using These 5 Effective Strategies

In response, LoDuca filed the full text of eight cases in federal court. But the problem only deepened, Castel said in a filing, because the texts were fictitious, citing what appeared to be “bogus judicial decisions with bogus quotes and bogus internal citations.”

The culprit, it would ultimately emerge, was ChatGPT. OpenAI’s popular chatbot had “hallucinated” — a term for when artificial intelligence systems simply invent false information — and spat out cases and arguments that were entirely fiction. It appeared that LoDuca and another attorney, Steven Schwartz, had used ChatGPT to generate the motions and the subsequent legal text.

READ ALSO:  Amazon hires former Microsoft product chief to oversee devices unit

Schwartz, an associate at the law firm of Levidow, Levidow & Oberman, told the court he had been the one tooling around on ChatGPT, and that LoDuca had “no role in performing the research in question,” nor “any knowledge of how said research was conducted.”

Opposing counsel and the judge had first realized that the cases didn’t exist, providing the involved attorneys an opportunity to admit to the error.

READ ALSO:  Interest rate worsens housing crisis, slows investment

LoDuca and his firm, though, seemed to double down on the use of ChatGPT, using it not just for the initially problematic filing but to generate false legal decisions when asked to provide them. Now, LoDuca and Schwartz may be facing judicial sanction, a move that could even lead to disbarment.

The motion from the defense was “replete with citations to non-existent cases,” according to a court filing.

“The Court is presented with an unprecedented circumstance,” Castel said. He set a hearing for June 8 when both LoDuca and Schwartz will be called to explain themselves. Neither attorney responded to CNBC’s request for comment.

Categories

Share This Article
Leave a comment