'AI Slop' in Court Filings: Lawyers Keep Citing Fake AI-Hallucinated Cases (indianexpress.com) 135
"According to court filings and interviews with lawyers and scholars, the legal profession in recent months has increasingly become a hotbed for AI blunders," reports the New York Times:
Earlier this year, a lawyer filed a motion in a Texas bankruptcy court that cited a 1985 case called Brasher v. Stewart. Only the case doesn't exist. Artificial intelligence had concocted that citation, along with 31 others. A judge blasted the lawyer in an opinion, referring him to the state bar's disciplinary committee and mandating six hours of A.I. training.
That filing was spotted by Robert Freund, a Los Angeles-based lawyer, who fed it to an online database that tracks legal A.I. misuse globally. Mr. Freund is part of a growing network of lawyers who track down A.I. abuses committed by their peers, collecting the most egregious examples and posting them online. The group hopes that by tracking down the A.I. slop, it can help draw attention to the problem and put an end to it... [C]ourts are starting to map out punishments of small fines and other discipline. The problem, though, keeps getting worse. That's why Damien Charlotin, a lawyer and researcher in France, started an online database in April to track it.
Initially he found three or four examples a month. Now he often receives that many in a day. Many lawyers... have helped him document 509 cases so far. They use legal tools like LexisNexis for notifications on keywords like "artificial intelligence," "fabricated cases" and "nonexistent cases." Some of the filings include fake quotes from real cases, or cite real cases that are irrelevant to their arguments. The legal vigilantes uncover them by finding judges' opinions scolding lawyers...
Court-ordered penalties "are not having a deterrent effect," said Freund, who has publicly flagged more than four dozen examples this year. "The proof is that it continues to happen."
That filing was spotted by Robert Freund, a Los Angeles-based lawyer, who fed it to an online database that tracks legal A.I. misuse globally. Mr. Freund is part of a growing network of lawyers who track down A.I. abuses committed by their peers, collecting the most egregious examples and posting them online. The group hopes that by tracking down the A.I. slop, it can help draw attention to the problem and put an end to it... [C]ourts are starting to map out punishments of small fines and other discipline. The problem, though, keeps getting worse. That's why Damien Charlotin, a lawyer and researcher in France, started an online database in April to track it.
Initially he found three or four examples a month. Now he often receives that many in a day. Many lawyers... have helped him document 509 cases so far. They use legal tools like LexisNexis for notifications on keywords like "artificial intelligence," "fabricated cases" and "nonexistent cases." Some of the filings include fake quotes from real cases, or cite real cases that are irrelevant to their arguments. The legal vigilantes uncover them by finding judges' opinions scolding lawyers...
Court-ordered penalties "are not having a deterrent effect," said Freund, who has publicly flagged more than four dozen examples this year. "The proof is that it continues to happen."
Make it stop quickly (Score:5, Interesting)
If a lawyer introduces hallucinated slop into a case that lawyer loses their license for at least one year. The lawyer must then retake the bar exam to regain their license. That'll make them think twice about using AI slop.
Re: Make it stop quickly (Score:2, Interesting)
When I was convicted of marijuana possession before the state legalized it, how come the prosecutors got away with writing the wrong amount of marijuana I was found with, which the judge caught but then simply corrected in court and moved on with the sentencing? Why shouldn't those prosecutors have been punished for hallucinating slop?
Re: Make it stop quickly (Score:5, Interesting)
I mistake is different from glaring lack of professional conduct.
Re: Make it stop quickly (Score:5, Insightful)
I mistake is different from glaring lack of professional conduct.
Using non-local AI in any way in court filings which are supposed to be confidential until filed is glaring lack of professional conduct right up front. Allowing AI hallucinations to get in to your court paperwork even once is the same. They should lose their license for one year the first time, five years the second time, and permanently the third.
Re: (Score:2)
I agree.
Re: (Score:3)
There is no excuse for submitting AI slop. When you file a court brief, you sign it indicating that you read it and it is as accurate as you can verify. You may quibble over details but you indicate everything you put in the file is factual.
Putting in fake case citations means you didn't read what you filed which means you violated your duty as a lawyer when you filed it.
Also - checking citations isn't hard. There's this tool called "Google" that you can spend 5 minutes with looking up citations. It doesn't
Re: Make it stop quickly (Score:2)
Re: (Score:2)
To be fair, lawyers do use external services that in theory leak a lot of information about their cases all the time, and have done for decades. Databases of case law are the obvious example. The searches give an insight into what the lawyer is thinking, what their likely arguments will be, things they may have overlooked.
Naturally those services offer confidentiality, the same as the phone company promises not to listen to the lawyer's call to their client, unless legally compelled to.
The question is, are
Re: (Score:2)
Are you calling for Luigi's release
Yes
Re: Make it stop quickly (Score:4, Insightful)
Can you explain how that's relevant?
It seems like you're fixated on the use or lack of use of AI, as opposed to the fact that a lawyer trusted a 3rd party service that is known for lacking accuracy and then presented citations that do not exist.
Ask yourself... what would this look like if AI wasn't involved. Would a lawyer be liable if they just made up a citation...? What about several? What about in an environment where a number of them had done it recently, and therefore there was a standing warning against doing it again?
Re: Make it stop quickly (Score:2)
Wikipedia is still seen as untrustworthy. On some subjects it's fine. On others it ain't. Jimmy is doing crimethink editing on behalf of Israel now.
Re: (Score:2)
Putting a wrong amount once is a careless mistake. If the prosecutor did it multiple times, I'd expect a pissed judge and wonder if the mistake is intentional.
Re: Make it stop quickly (Score:2)
How do you know the judge sees all their mistakes before other judges? And in the article, how do you know each lawyer didn't just make one careless mistake when using AI?
Re: (Score:2)
Yes, of course the judge would see the same prosecutor make mistakes. There aren't that many prosecutors and that many judges in a district.
AI is probably great for summing up cases and writing first drafts. The rest is the lawyer's actual job. If the lawyer typos a year or spells a name wrong, it's the same level of mistake that you dealt with, but it's not malpractice.
Re: Make it stop quickly (Score:2)
In a rural conservative district, is it conceivable that a judge would let prosecutors get away with lots of mistakes when it comes to drug infractions?
Re: (Score:2)
So the judge and the prosecutor have it in for you with your tiny drug charge? You just aren't that important, never have been, and never will be. How long did you serve? Hard time with a rapist named Biff?
Re: Make it stop quickly (Score:2)
Do lawyers make mistakes all the time that you dismiss in your zeal to demonize anything AI?
Re: Make it stop quickly (Score:2)
Are both examples of lazy prosecutors, but in one case you get outraged and in the other you find excuses why it doesn't matter that the prosecutor was sloppy with facts?
Re: Make it stop quickly (Score:3)
Re: (Score:3)
What if my judge had thrown out the prosecutor's case due to their negligence in lying on court documents about the amount of marijuana I was in possession of? Why didn't he? What if AI mis-citations don't misrepresent the spirit of the law? Do you have evidence or just strong feelings that they do?
First of all, lying is a strong word. You can't know there was any intention behind the error.
But if the judge had thrown out your case, then a person who committed an act that was against the law would have gone free. Which isn't a good thing, regardless of what was or wasn't legalized after the fact and regardless of how you and I feel about that act. (For the record I couldn't possibly care less about pot possession or use.)
Why didn't he? Presumably because clerical mistakes aren't sufficient caus
Re: Make it stop quickly (Score:2)
Because it's not the same.
Humans make honest mistakes all the time, we fix them and move on, and we give benefit of the doubt. If we can prove that the mistake isn't an honest one, then it's different, but you haven't mentioned anything (here or below) that suggests otherwise in the case you mention.
Using generative AI is making a conscious decision to use a technology that makes shit up. There should be no tolerance for that.
Re:Make it stop quickly (Score:5, Informative)
Technically Judges cannot do that directly. Instead the procedure is:
1) Report to Bar
2) Have a Hearing by the Bar
3) The bar can decide to take their license for X amount of time.
But I do agree that is what the Judges should be doing.
The judges can however hold anyone in contempt of court for any reason at any time. Do not even have to be in court. You can appeal it, but as long as the Judge was somewhere near reasonable, you will not succeed.
If however the Judge does something like hold the Umpire at his kid's baseball game in Contempt, then yes you will almost certainly win the appeal.
Re: (Score:2)
Re: Make it stop quickly (Score:2)
Re: Make it stop quickly (Score:2)
Re: (Score:2)
Not lose their license. The case should be immediately dismissed in favor of the other side.
Let the client sue the lawyer for damages. That will fix the issue far faster than even disbarment.
Re: (Score:2)
Yes, I like this too.
And this will go on and on. Until? (Score:5, Insightful)
Re: (Score:3)
Re: (Score:3)
> No need for all that. Either "Judgement is for the other side" or "Case dismissed." Clears the docket, and slows down these kinds of submissions until they're at least doublechecked.
Interesting. I think you've changed my mind about this.
Economic incentives are probably the way to go.
Re: And this will go on and on. Until? (Score:5, Informative)
It doesn't seem fair the clients should suffer the burden of bad representation. Even in civil cases. Let alone criminal.
Economic incentive needs to target the right party.
Re: (Score:3)
It doesn't seem fair the clients should suffer the burden of bad representation. Even in civil cases. Let alone criminal.
Economic incentive needs to target the right party.
You also don't want them to be selecting for poor representation because they receive the outcome they desire - pushing off punishments.
Re: And this will go on and on. Until? (Score:2)
Alright so I don't know much about the US legal system. Imagine my lawyer uses genAI, gets caught for "hallucinations" and I lose the case. Can I then sue my lawyer for that? How likely is it that some other lawyer accepts to represent me in that case?
Re: And this will go on and on. Until? (Score:2)
IANAL. I'm guessing you probably can sue, if you have money. That won't get you very far unless the other side has money. Imagine it's a public defender, and you have automatically lost a death penalty case because they used AI.
Re: (Score:2)
I'm just glad that cops tend to be too computer illiterate to start using chatbots to file their cases (so far). There's at least some slim chance the Bar may rule against one of its own, but the Thin Blue Line routinely covers up crimes as serious as murder and drug running so there'd be no hope from that direction.
Re: (Score:2)
There are examples where rulings have been overturned or trials had to be redone due to someone having ineffective counsel.
Re: (Score:2)
The law has accommodated in the past for parties egregiously/negligently screwing up, prior to the introduction/assistance of technology.
Re: (Score:2)
Typically we don't destroy an entire firm for the misconduct of one employee, unless it's so extreme that it justifies screwing all their other clients. Imagine if your case was headed to court and your lawyer said their firm had been wiped out by another employee using AI, so you need to find another lawyer and hope the court is willing to accommodate the delay. Even if the court is, re-doing much of the process, document exchange, and so on will take a lot of time and create more expense, that you might e
I suspect they've always done this (Score:4, Insightful)
But I would bet money if you did a exhaustive analysis of court filings you would find plenty of made up citations that nobody ever looked into.
Then again lawyers are famous for their honesty and decency so maybe I shouldn't make assumptions.
Re:I suspect they've always done this (Score:5, Insightful)
Re: (Score:3)
Which now bites them in the ass. This is a classical case of greedy dumb assholes trying to do things "cheaper than possible".
Re: (Score:2)
A distinct shift in who messed up. The AI can't be blamed.
Re: (Score:2)
And what does that say about the trust we should have in today's justice system and government in general? Lots of lawyers in government!
Re: (Score:2)
I mean, you can't trust the politicians and have to hope the courts keep them in check...
Re: (Score:2)
Citizens United.
'Nuff said.
Re: (Score:2)
Re: (Score:2)
Fake references have a long and disgraceful history. Cram 30 or 40 citations into a filing and the judge might look at the first few and assume the rest are more of the same. IIRC this has occurred even in case filed before the Supreme Court, where the last citations were actually cases which pointed in the opposite direction that the lawyer wanted but no one bothered checking for quite a long time.
Lazy People == AI Slop (Score:2)
This means you
Debugging LLM (Score:3)
This gets into the issue of how to debug an LLM?
Since this keeps happening, it seems like when a model is trained for legal cases, it should be done in such a way that the underlying source material is tied to the case.
For example, if the model reads about Johnson v. Smith, 1972 in "Legal Shit today, issue 175", then it should be possible to query the LLM and ask it what the source material is for that case, and get back "Legal Shit Today, issue 175". This feature would at least start to allow you know how it is coming up with hallucinations.
With that in place, the sequence of events then becomes:
Write me a court filing for issue (whatever)
Give me the list of cases cited in the doc you just wrote
Show me the sources for those cases cited
Manually research the sources, verify each case cited
Re: (Score:2)
Manually research the sources, verify each case cited
Clearly this not even even being done by an automated tool, let alone a human. An LLM which is given access to a database of actual cases could reasonably be successful at checking whether the cased cited even exist which isn't being checked now!
Re: (Score:2)
Well that's kind of my point. Other than, I was proposing human do it, rather than having an AI do it.
Re: (Score:2)
I'm not against a human doing it, I just don't get why they're not having the software do it, when it's feasible for it to do it.
Re: (Score:2)
Ah, no. LLMs cannot do fact-checking. They hallucinate even with full access to verified facts.
Re: Debugging LLM (Score:2)
You can however feed the output of one LLM to another LLM, and ask it to fact check. I have done so. It is worthwhile. I think in the case of completely made up cases, the second pass (or third) is very likely to flag it due to no correlation in its database.
For the logic in legal arguments, though, I would not expect it to help.
Re: (Score:3)
You cannot "debug" an LLM. Not possible. That is one of the reasons why LLM use requires a high level of on-target expertise and a lot of care.
Re: Debugging LLM (Score:2)
Do you remind yourself of Herbert Simon proclaiming in the 1960s that neural networks were a dead end because perceptrons couldn't do XOR?
"Simon was right in one sense: early perceptrons couldnâ(TM)t scale to general intelligence.
But he was wrong in thinking the flaw was essential to connectionism.
Once computational power, training methods, and deeper architectures arrived, neural networks surpassed the very symbolic reasoning systems Simon pioneered."
Re: (Score:2)
What a stupid comment that entirely misses the point.
Re: Debugging LLM (Score:2)
Is the point that you are making a lot of assertions why LLMs are unworkable just as Simon did about neural networks?
Re: (Score:2)
No? Why would you even think that except if you just wanted to throw crap?
Here in the real world, a perceptron is a very simple thing and an LLM very much is not. Are you unaware of that?
Re: Debugging LLM (Score:2)
---
When the perceptron was invented in the 1950s, it wasnâ(TM)t seen as a âoesimple thing" at all. Frank Rosenblatt and the U.S. Navy promoted it as a major step toward machine intelligence. A New York Times headline in July 1958 read âoeElectronic âBrainâ(TM) Teaches Itselfâ (NYT, July 13 1958: [https://www.nytimes.com/1958/07/13/archives/electronic-brain-teaches-itself.html](https://www.nytimes.com/1958/07/13/archives/electronic-brain-teaches-itself.html)). Rosenblatt predict
Re: (Score:2)
Re: (Score:2, Insightful)
It just means the penalties are insufficient.
If the penalty for speeding were life imprisonment, you would see a huge, but not total, reduction in speeding.
Assuming the law were enforced. If you don't enforce the law, the penalty is only theoretical.
Re: (Score:2)
It just means the penalties are insufficient.
That is the no-clue cave-man answer. All research shows that increased penalties have no positive effect, but make the problem worse.
Re: (Score:2)
Well, speaking for myself, I can tell you that if I'm driving, I'm probably speeding. If the penalty were life imprisonment, that would certainly put an end to my speeding.
I don't think I'm particularly unique in that regard, but sure, anything is possible.
Re: Idiotic statement (Score:2)
Have you heard of the forbidden fruit effect? Do you think prohibition stopped drugs?
Are you certain that the harm of a few mis-citations merits the kind of draconian response you're calling for, and will the unintended consequence be that the people elect AI Trumps to undo your ridiculously over-the-top punishments for using AI?
Re: (Score:2)
I am not speaking for myself. I am speaking for solid results from criminology. You think you are smarter and know more than tons of researchers that have studied this for a long, long time?
Re: (Score:3)
Re: Idiotic statement (Score:2)
If you have a democratic society why wouldn't people vote to overturn stupid ideas like impounding cars for speeding?
Re: Idiotic statement (Score:2)
Why wouldn't you vote to end such a ridiculous penalty? Remember when we the people undid drug prohibition in many states?
Re: (Score:2)
The idea that harsh punishments for crimes will lead to people not committing crimes in fear of those punishments is not new. In Victorian times you had the 'death penalty for everything' approach to sentencing in many countries. It's been changed in large part because it didn't work. People ended up stealing because of poverty and once they were liable to be hanged, many figured, the law already can't throw any more at them, so they continued robbing and killing and became outlaws.
Maybe you won't become a
Re: (Score:3)
All research shows that increased penalties have no positive effect, but make the problem worse.
It also shows that if the penalty is insufficient then they have no positive effect. A fine that people with a lot of money can easily afford is just a prohibition which only applies to the poor, with a license fee. Look to speeding tickets which scale with income for a fair model.
Re: (Score:2)
Think about the logic you are applying here. It makes absolutely no sense.
Re: Idiotic statement (Score:2)
I'm sorry you didn't follow it, and then felt a need to say so. Which part do you find confusing? The part where it's not actually a deterrent if it's not actually a penalty?
Obligatory (Score:2)
It Never Ceases to Amaze Me (Score:2)
Re:It Never Ceases to Amaze Me (Score:5, Insightful)
The thing is that LLMs cannot fact check. Apparently the users of LLMs are, in this case, too lazy or to dumb (or both) to do it themselves.
There is indications that for many things, LLMs actually decrease efficiency and using them is a costly mistake. This ("better search") is supposedly one of the few areas where they save time. But fact-checking LLM results actually requires more skill and competence than generating the results manually and it MUST be done for results of reasonable quality. Apparently, that little problem is still not common knowledge.
Re: (Score:3)
I do not "think so". I am just referring to research results. You, on the other hand, have nothing. LLMs cannot fact-check, period.
Robotics just uses the fact that you can train whatever you use pretty well and you can use _other_ systems there to install guardrails and these other systems come with physics models and can do (limited) fact checking in there. These other systems are not LLMs though. LLMs cannot fact-check in robotics either.
Re: It Never Ceases to Amaze Me (Score:2)
Is the irony lost on you that you provide no citations to support your probable hallucinations about fact checking made-up citations?
Disbar them (Score:2)
[C]ourts are starting to map out punishments of small fines and other discipline [....] Court-ordered penalties "are not having a deterrent effect," said Freund, who has publicly flagged more than four dozen examples this year. "The proof is that it continues to happen."
Disbar them then - they're not doing their fucking job properly, so why should they get called "lawyers/attorneys" then?
And if they do want to be a lawyer/attorney again, then I guess they'll have to re-enroll, study, and take the bar exam again. Without using any A.I.
Imagine if a plumber relied on AI to fix your massive water leak. Would you seriously pay for that?!?
Re: (Score:2)
If it fixed the leak? Why not? Especially if it enabled them to fix it faster and/or cheaper.
Just FYI, AI is in use in the construction trades already, most people aren't aware of that. For your example a draftsman can feed the plans of a building into an adequately trained system and map out the most efficient routing for plumbing and cabling. AI is operating excavators, scheduling contractors, driving inspection robots, recognizing bad concrete pours from drone images, and the list keeps growing. In
Re: (Score:2)
map out the most efficient routing for plumbing and cabling.
That's a toy program for finding shortest distance. It won't understand specifications or building code. It won't understand working with multiple trades. It definitely won't understand any style other than post-modern brutalist.
Re: (Score:2)
It won't understand specifications or building code.
It most certainly does. I did say "adequately trained".
Re: (Score:2)
It
So no program name, just "it"?
trained
As in LLM and not Expert System?
Specifications are basically unique to each building.
Training an LLM on specifications will only make it hallucinate more than usual.
Re: (Score:3)
Treat it as if they wrote it themselves, which means fraud, falsifying documents, lying under oath.. etc.
"Science" has the same problem, thank you RFKjr... (Score:4, Insightful)
RFKjr's administration have been using AI to generate justifications for policies that all are hitting exactly the same problems:
* AI is inventing studies that never existed
* AI is using quotes from real studies that aren't in the studies
* AI is generating summaries of studies that are the opposite of what the study itself actually concluded
and he's referencing these AI generated summaries in congressional hearings.
Re: (Score:2)
The funny thing is that all of that would be very easy to verify. But LLMs are completely unable to verify. Such a great tool...
Re: (Score:2)
LLMs are completely unable to verify.
That's an exaggeration. You can give a LLM access to real things and they can use those real things to verify. I just flatly do not understand why they are not. It wouldn't make them infallible, but it would go a huge way towards improving the situation, and they are clearly not doing it. They could also use non-AI software tools to check up on the AI output. I'd bet that you could even use a plagiarism detection tool for this purpose with little to no modification, but I'd also bet this kind of tool alread
Re: (Score:2)
No, they cannot. LLMs cannot do logical reasoning and that is a requirement for any type of fact-checking or verification. All they can do is correlations and that is not enough.
While using non-LLM AI or non-AI tools is a possibility for fact-checking (in simple cases as the one here), you overlook that basically the only advantage of LLMs is that they are comparatively easy to create. Creating these "other tools" that would be needed here would be a lot more expensive and a lot more work than creating gene
Re: (Score:2)
All they can do is correlations and that is not enough.
It's enough to determine whether a citation even exists, which is what this story is about.
Re: (Score:2)
And that is a problem that is actually pretty hard and expensive to solve. For software. An experienced and smart human can do it easily. A machine cannot.
Prison time (Score:4, Insightful)
At the rates these fuckers charge, this is completely unacceptable and should be regarded as fraud.
Here's the database (Score:5, Informative)
The article above mentions a database used to track these AI hallucinations in legal cases, but fails to link or name it. Here it is:
https://www.damiencharlotin.co... [damiencharlotin.com]
The database is well-structured for further investigation. One interesting point is that the majority of transgressions were committed by pro se parties, with lawyers a close second.
Re: (Score:2)
One interesting point is that the majority of transgressions were committed by pro se parties...
!!!
That's a key point. Most of the AI generated slop is done by people representing themselves.
https://www.law.cornell.edu/we... [cornell.edu]
How do we still have an AI bubble? (Score:5, Insightful)
High profile hallucinations are in the news every few weeks. The MAHA report cited medical research that never happened.
It's been a constant problem since ChatGPT was released. There is no solution on the horizon.
And yet, there are still people who are betting that this technology will be an economic game changer.
Re: (Score:2)
It is a game changer, just not the way it's being used. The accellerating advances of robotics are almost all from the use of AI trained in specific tasks. ChatGPT couldn't operate a robot, but neither could the AI that runs Boston Robotics' Spot converse with you.
Re: (Score:2)
There is no solution on the horizon.
The solution is to do their own work! Or at least check it! This use of AI is like when people just Google for something and copy/paste the first hit.
And yet, there are still people who are betting that this technology will be an economic game changer.
What the news doesn't show is the millions of people using AI successfully every day. We know that every day people incorrectly use hammers, wrenches, screwdrivers, drills, cars, and guns. Yet they are not declared useless. AI is a tool, and in the hands of a someone with genuine interest in using the tool appropriately, it is a useful one. The best path
Knock-on effects? (Score:2)
It seems certain that plenty of these AI-hallucinated precedents are being entirely missed in court cases. So what happens if a case in which a hallucinated precedent figures becomes a precedent in and of itself?
Could a fake precedent become sane-washed in this manner to the extent that it becomes precedent? I just asked my wife, who is in a position to know, and she confirms that yes, it is entirely possible. So we're not just looking at AI fucking up one decision incorrectly. We're in danger of AI actuall
Search Engines on Steroids (Score:2)
Lazy frakkers. They never heard of proofreading? I understand using AI to get ideas, but to turn over your professional reputation to one is pathetic. It's on the same level as plagiarism.
Citations (Score:5, Insightful)
The real solution would be a citation system.
Something like LexisNexus has every court case that happens in the country.
So... why not have an official version of that, tied in with the official court transcripts and when you cite a case, you need to give that citation number from the official database. If you're citing only a few lines, you link to those few lines.
You wouldn't be able to cite a non-existent case, at best the case you cite wouldn't match what you claim it does, and with individual statement citations (HTML literally does it already), you could prove in one click that that series of words actually appears in that cited case.
You want to stop this? Then open-source the law instead of hiding it behind stupendously expensive private commercial services like LexisNexus.
Re: (Score:2)
It's akin to the cops' Thin Blue Line, the judge is also a lawyer, he doesn't want to upset his coworkers.