Texas judge says no AI in courtroom unless lawyers certify it was verified by human

Judge cited AI hallucinations and bias in that decision

A federal judge in Texas is requiring lawyers in cases before him to certify that they did not use artificial intelligence technology to draft filings without a human checking their accuracy first. 

U.S. District Judge Brantley Starr of the Northern District of Texas issued the requirement on Tuesday.

Starr said in an interview Wednesday that he devised such a requirement to warn lawyers that AI tools can create fake cases and that he may sanction them if they rely on AI-generated information without first verifying it.

"We're at least putting lawyers on notice, who might not otherwise be on notice, that they can't just trust those databases. They've got to actually verify it themselves through a traditional database," he said.

MITIGATING 'RISK OF EXTINCTION FROM AI' SHOULD BE GLOBAL PRIORITY, INDUSTRY LEADERS SAY

A general view of the Earle Cabell Federal Building

A general view of the Earle Cabell Federal Building on June 17, 2019, in Dallas. (Photo by Ron Jenkins/Getty Images / Getty Images)

Starr explained that he began drafting the mandate while attending a panel on artificial intelligence at a conference hosted by the 5th Circuit U.S. Court of Appeals.

The judge said he considered banning the use of AI in his courtroom altogether, but had decided not to do so after conversations with UCLA School of Law professor Eugene Volokh.

Starr also noted that he and his staff will avoid using AI in their work altogether.

ChatGPT website displayed on a laptop

The ChatGPT website displayed on a laptop screen in is seen in this illustration photo taken in Krakow, Poland, on April 11, 2023. (Photo by Jakub Porzycki/NurPhoto via Getty Images / Getty Images)

"I don't want anyone to think that there's an algorithm out there that is deciding their case," Starr said.

GET FOX BUSINESS ON THE GO BY CLICKING HERE

In a notice on the district court's website, the order said all attorneys must attest that either no portion of the filing was drafted by generative artificial intelligence – like OpenAI's ChatGPT or Google Bard – or that any language drafted by generative artificial intelligence was checked for accuracy, "using print reporters or traditional legal databases, by a human being."

BARD from Google

In this photo illustration, the Google Bard logo is seen displayed on a smartphone. (Photo Illustration by Rafael Henrique/SOPA Images/LightRocket via Getty Images / Getty Images)

The statement said that while such platforms are "incredibly powerful," platforms in their current states are prone to hallucinations. In addition, reliability and bias are other issues to consider. 

CLICK HERE TO READ MORE ON FOX BUSINESS

"Unbound by any sense of duty, honor, or justice, such programs act according to computer code rather than conviction, based on programming rather than principle," the notice said. 

Reuters contributed to this report.