
OpenAI Says Models Programmed To Make Stuff Up Instead of Admitting Ignorance (theregister.com) 90
AI models often produce false outputs, or "hallucinations." Now OpenAI has admitted they may result from fundamental mistakes it makes when training its models. The Register: The admission came in a paper [PDF] published in early September, titled "Why Language Models Hallucinate," and penned by three OpenAI researchers and Santosh Vempala, a distinguished professor of computer science at Georgia Institute of Technology. It concludes that "the majority of mainstream evaluations reward hallucinatory behavior."
The fundamental problem is that AI models are trained to reward guesswork, rather than the correct answer. Guessing might produce a superficially suitable answer. Telling users your AI can't find an answer is less satisfying. As a test case, the team tried to get an OpenAI bot to report the birthday of one of the paper's authors, OpenAI research scientist Adam Tauman Kalai. It produced three incorrect results because the trainers taught the engine to return an answer, rather than admit ignorance. "Over thousands of test questions, the guessing model ends up looking better on scoreboards than a careful model that admits uncertainty," OpenAI admitted in a blog post accompanying the release.
The fundamental problem is that AI models are trained to reward guesswork, rather than the correct answer. Guessing might produce a superficially suitable answer. Telling users your AI can't find an answer is less satisfying. As a test case, the team tried to get an OpenAI bot to report the birthday of one of the paper's authors, OpenAI research scientist Adam Tauman Kalai. It produced three incorrect results because the trainers taught the engine to return an answer, rather than admit ignorance. "Over thousands of test questions, the guessing model ends up looking better on scoreboards than a careful model that admits uncertainty," OpenAI admitted in a blog post accompanying the release.
No shit (Score:2)
Watson.
Re:No shit (Score:4, Interesting)
Re: (Score:2)
Concurrence in general but I think the underlying problem is that it doesn't ask for clarification when it should. Even worse than the verbosity thing.
Re: (Score:2)
Re: (Score:2)
at first I thought the verbosity thing was kind of cool. Now I wish it would give me more precise answers. and.... (me complaining), when I feed it code, I have to specify to change one small part of it, but in general, it seems to want to modify all of it, and in many cases it inserts bugs. So I learned, only give it small snippets.
Often the verbosity of the response can be tailored via the appropriate additional qualifying words in the prompt. So, couldn't this be transparently baked into a personalized interface? That is, allow the user to specify the preferred verbosity as part of personalized settings and silently add the appropriate additional words to the prompt each time. The user can still explicitly ask for more verbosity when desired.
Anecdotal experiences with gen AIs (Score:2)
Well I can report that just telling the genAIs to be less verbose doesn't help, though I recall DeepSeek promising to try.
Re: No shit (Score:2)
Have you ever tried to instruct it to ask you for clarifying questions before providing an answer?
More anecdotal experiences with gen AIs (Score:2)
Yes, I have tried to encourage the genAIs to ask questions a number of times in various ways, but with no significant success that I can recall.
I can definitely recall the results of my last paired experiments. I prepared a short software specification and gave that text to DeepSeek and ChatGPT.
The DeepSeek result was better in terms of how I described the desired appearance, but one of the four results was wrong. DeepSeek had clearly misunderstood that part of the problem and did NOT ask for clarification.
Re: (Score:1)
The potential energy itself is lost because that stored charge in the small capacitor pushes charge out to spread evenly to the initially uncharged capacitor. The energy itself was used up in the electrons pushing away from eachother. Even with ideal capacitors and zero resistance - the energy need not be lost as heat.
Sort of like letting a compressed spring go. Even if the spring did no working lifting or pushing against any other object, it still lost potential energy springing itself back to rest.
Re: (Score:2)
Re: (Score:1)
The lumped circuit element model with ideal wires, capacitors, and instantaneous effects is not reality, and ltspice doesn't capture the whole picture. Simulating what would happen with electrons in a capacitor with superconducting plates and superconducting wires in between, would show the effects of magnetic fields and electric fields and EM radiation. In a zero-resistance environment, it would likely oscillate (due to the changing magnetic field and the wires having intrinsic inductance) and dissipate so
Re: (Score:2)
I recently gave AI's a paradox. When an ideal 10uF cap is charged to 100V, it has x charge, when an uncharged cap of 100uF is placed in parallel...
Is the 100uF cap also ideal?
Ideal Capacitors not the Problem (Score:2)
Re: (Score:2)
I though his question was more than clear enough TBH.
You're being more than needlessly pedantic about what parallel means here to the point of being wrong. It doesn't imply that it's connected to something else. If I have an array caps connected in parallel and disconnect that array from the supply, they have not ceased to be in parallel. Parallel simply implies that you have a bunch of components where all their terminal A's are connected together and all the terminal B's are connected together.
Once you kn
Re: (Score:2)
You're being more than needlessly pedantic
No, I'm being appropriately pedantic for a physics question. When you have a capacitor charged to 100V and another is connected in parallel to it the usual implication is that this too will be charged to 100V since connecting something in parallel implies the same pd across the device. If the capacitor were already disconnected from the external power then parallel and series have no meaning since they would both be the same as your circuit now consists of just two capacitors.
Thus, in specifying paralle
Re: (Score:2)
If the capacitor were already disconnected from the external power then parallel and series have no meaning since they would both be the same as your circuit now consists of just two capacitors.
Well, no, because the person clearly said how they were connected. Name the terminals of a capacitor A and B and indicate which capacitor by 1 and 2. A paralleled pair of capacitors has A1 and A2 connected and B1 and B2 connected. You now have a 1 port network with A1/B1 as the ports.
A series pair has A1 and B2 disc
Re: (Score:2)
Re: (Score:2)
The literal definition of a parallel circuit is one where the circuit divides and the current is split between two components - look it up.
You know I actually reached behind me and grabbed Horowitz & Hill off the shelf just because.
Looks like someone has h4x0rized it so you can look too:
https://kolegite.com/EE_librar... [kolegite.com]
Page 2, Figure 1.1, "parallel connection". See the lack of an EMF in that diagram?
When the OP said "capacitors in parallel", that's exactly what he means. Those wires on the left and righ
Read the Text (Score:2)
Page 2, Figure 1.1, "parallel connection". See the lack of an EMF in that diagram?
Yes, now see the text directly under the diagram which says, and I quote, "Things hooked in parallel (Figure 1.1) have the same voltage across them.". So no the source of the EMF is not shown but it is clearly there as the text underneath states. Thank you for proving my point.
OKey dokey, since you keep dodging this question I'll ask again.
The line you quoted answers the question you asked: if the current does not divide between two or more paths then the devices are not connected in parallel. Apply that to your situation: if there is a current and it splits to pass t
Re: (Score:2)
"Things hooked in parallel (Figure 1.1) have the same voltage across them."
Ah you mean the entire premise of the OP's question?
So no the source of the EMF is not shown but it is clearly there as the text underneath states. Thank you for proving my point.
At this point I have to wonder if you know how capacitors work. If those two floating wires are unconnected to an EMF, the two capacitors will still have the same voltage across them, due to whatever charge/energy is stored in them.
You don't need an externa
Re: (Score:2)
If those two floating wires are unconnected to an EMF, the two capacitors will still have the same voltage across them, due to whatever charge/energy is stored in them.
There are two problems with this. First what happens if the components are resistors? Is this some new rule you have invented that only applies to capacitors while resistors in the exact same arrangement will not be connected in parallel just because they cannot generate their own EMF? Next this argument seems to distinguish whether something is in parallel or series based on the addition of two unconnected wires. If I remove those two, unconnected wires are you trying to tell me that this will convert the
Re: (Score:2)
Well, shit. My phone are my reply which was quite long. So sorry my second go will be worse.
The EMF requirement is yours not mine. I don't see why an emf of 0 is a problem. Just a special case of complex impedances. But given your requirement capacitors fit.
.
I've also never heard the term "series loop" before, and that only holds for a pair. What would you call it work 3, 4,5 or more capacitors connected how I specified as parallel?
When you say"no it only behaves that way", that's what I was talking about.
Re: (Score:1)
Re: (Score:2)
I recently gave AI's a paradox. When an ideal 10uF cap is charged to 100V, it has x charge, when an uncharged cap of 100uF is placed in parallel, the voltage goes to about 9.1V and the total energy is about 0.1x. If energy is neither destroyed nor created, where did the energy go? The AI kind of forgets that I specified an "ideal" capacitor, and makes shit up. If a human can explain this to me, I am all ears. Please keep in mind: "Ideal capacitor", and "equations show this", don't make shit up.
Consider the moment just before the two caps are connected to each other. One is charged to a nonzero voltage and the other is at zero. As the conductor attached to one capacitor top plate approaches the conductor attached to the other capacitor top plate (the bottom plates already sharing a common connection), the E field between them will at some point be high enough to exceed the breakdown field of the medium in between (even if it is free space). Then an arc occurs. That arc is observable, maybe even vi
Re: (Score:2)
Now what AI does with this is a completely different question.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
And yes, the AI explanation is not based on the basic formulas. That is actually the problems. Humans can sanity check an answer looking for conservation of energy. AI does not take this step. For these types of problems, a little recursion with some simple application of basic laws of physics sh
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:1)
The energy present in the first capacitor is 1/2 * 10uF * 100V^2 = 0.05 Joules.
The energy present in the two after stabilization is 1/2 * 110uF * 9.09V^2 = 0.00455 Joules.
So, 90% of the energy is lost during the connection.
The "two-capacitor paradox" is well documented. Conservation of energy must count all possible types of energy (not just stored charge), including heat, photons, and electromagnetic radiation during the energy transfer from one capacitor to the other. Even if you have ideal capacitors and
Re: (Score:1)
the product of voltage (V and capacitance (C) gives you charge (Q), not energy. Q=C*V ; E=1/2*C*V^2.
The energy present in the first capacitor is 1/2 * 10uF * 100V^2 = 0.05 Joules.
The energy present in the two after stabilization is 1/2 * 110uF * 9.09V^2 = 0.00455 Joules.
So, 90% of the energy is lost during the connection.
The "two-capacitor paradox" is a well documented and well explained thought problem. At the moment of connection, the current flow is extremely high, producing a burst of EM radiation (e
Re: (Score:1)
The "two-capacitor paradox" is well documented. The first problem is a dimensional error, comparing charge (e.g. coulombs) and energy (e.g. joules). But another problem is that conservation of energy must count all possible types of energy (not just stored charge), including heat, photons, and electromagnetic radiation during the energy transfer from one capacitor to the other. Even if you have ideal capacitors and zero resistance, with infinite current for an infinitesimal time at the moment of connection
Re: (Score:2)
Re: (Score:1)
Any of flipk's 3 similar responses does a pretty good job of explaining what is going on: Losing energy to heat and/or EM radiation. Its not really a paradox; just some difficulty fully accounting for everything that affects the total energy of the system, as opposed to conservation of charge approach where energy can basically be ignored.
It reminds me of the first day of the first physics course I took as a senior in high school. The teacher gave us a pre-test (of multiple-choice questions) as part of a
Re: (Score:2)
The circuit itself has some of the energy.
Re: (Score:2)
Watson.
Speaking of Watson, Slate recently ran a retrospective article on IBM's Watson playing Jeopardy!. Coincidentally, it also mentioned when Watson made up the response that Toronto was a U.S. city:
https://slate.com/culture/2025... [slate.com]
Re: (Score:2)
aka Modeling their developers.
Human do the same (Score:1)
Wrong explanation (Score:5, Informative)
They make shit up because they have no meta-cognition and don't know any better.
Re:Wrong explanation (Score:5, Insightful)
Re: (Score:1)
This is a great time to remind ourselves that a LLM is just a fancy autocomplete engine.
Well, sure, in the same way a 747 is just a fancy mechanical bird. Which is to say, yes, but no.
Re: (Score:2)
If a bunch of people are marketing a 747 as a vehicle like some alien ship that can pause in mid air, go side to side, and backwards and underwater and into space, then it's worth pointing out it's closer to being a mechanical bird than an alien ship.
Re: (Score:2)
Hence why your AI has a 'seed' value.
Re: (Score:2)
Using the word 'programming' in the headline is a stretch to the point of tearing
Re: (Score:2)
the engine to return an answer, rather than admit ignorance
So it's doing the same thing as any guy would then?
Re: (Score:2)
They make shit up because they have no meta-cognition and don't know any better.
The whole point of this article is to claim that's not true. (I emphasize that we are not talking about the *sense* of knowing things but rather about an information flow that contains redundancy and can differentiate between justified knowledge and anything else.)
I've been saying there was something wrong in the architecture of these models for ages. If it just turned out the training was just subtly wrong, that would be disappointing. (Have we needlessly been on the wrong path?)
This is good research (Score:2)
Instead of claiming that their work is perfect, they look for problems to fix
All companies should do this
Unfortunately, some companies not only try to keep problems secret, but also try to punish those who expose the problems
And that's why I cancelled (Score:5, Insightful)
Re: (Score:2)
rtfm :( (Score:1)
Re: (Score:1)
Yes, but if you could make a little AI assistant that had knowledge of all your manuals and other important reference material that had already been vetted, would it not be useful and faster to just ask your assistant for that information? It sure sounds like it would be.
That's actually where I see these LLM type systems working best. You train them on well vetted material for your specific business, it's policies, etc and the AI can more or less be the very easy to use gateway to that information for your
Re: rtfm :( (Score:1)
Re: (Score:2)
It sounds like you want a search engine. You want to be able to index your own documents with a search engine, right?
Do you really need to have a conversation with the search engine in which it can misunderstand what you're trying to say, based on some association it learned from reddit posts?
Technically you can do what you're asking, but it's less efficient and more prone to error than just using a search engine... which you can also set up locally yourself.
Echo chamber effect? (Score:3)
It seems to me that LLM hallucinations which remain undetected in the short term, also give rise to false data which both reinforce the hallucination of the model in question and pollute other LLMs with false data. In other words, does one LLM, unchallenged when it makes shit up, adopt its own answer as fact and then corrupt other models which aren't 'guessing' but are simply propagating a lie unknowingly? Do LLMs treat each other's data as authoritative?
Please excuse the anthropomorphism in what I just wrote - it seems to be the most efficient way of conveying the concepts.
Re: (Score:2)
You mean is it eating it's own dog food? At this point, probably. A lot of places are using AI for "news" these days. If all that is scrapped, not to mention all the AI produced video. I could definitely see AI learning misinformation that was generated by a previous iteration of itself. The irony. Also very fortellable.
Re: (Score:2)
taking anthormorphism into account... absolutely yes.
It's also deeper than that though... seems like you're thinking of ideas as the atomic components that LLMs use, but they actually use word-parts and associations between them. Model collapse can be much more uh... interesting than just having them constantly come up with false statements.
Re: (Score:2)
taking anthormorphism into account... absolutely yes.
It's also deeper than that though... seems like you're thinking of ideas as the atomic components that LLMs use, but they actually use word-parts and associations between them.
I am profoundly a 'word' person, and a LOT of my thinking takes place explicitly in the language domain. Lately - prompted by considering LLMs and their behaviour, as well as the differences between humans and other animals - I've been playing with the notion that human ideas are at least in large part "word-parts and associations between them".
Re: (Score:2)
We... do use those methods, but not exclusively?
Some functions are not purely semantic. Given how language models actually rely on this perception to work, it's sort of dangerous to commit too hard to that line of thinking IMO.
That's the most annoying thing about AI (Score:1)
That's hands down the most annoying thing about AI. The one thing I miss from all models are more frequent "I don't know" responses. Knowing that the model doesn't know is a hugely valuable piece of information. And far more useful than hallucinations, half-truth that doesn't really fit the question, etc.
Generative vs Factual (Score:5, Insightful)
I think the problem is definition itself. Generative AI. If they need to generate an answer, then good chances are it'll end up whatever the model believes to be correct, statistically speaking.
But on the internet actual facts and answers are rare. Most of help-me threads are 99.99% crap and one or two people providing an actual, helpful response but those few units drown among all the other crap because hey - statistics don't favour truths.
Kind of like in democracy. Two uneducated dropouts have more power than one university lecturer.
But I digress...
If AI models were to return facts, we'd call them search agents, search engines...
Oh, wait.
Re: (Score:2)
Did you read the paper?
The paper is about whether we reinforce it to answer correctly (whatever that means) vs it simply saying "I don't know".
Most LLMs are rewarded when they give a correct answer. Say: 1 for correct, and 0 for incorrect.
So they are rewarded to always try an answer even when they aren't certain. The paper suggests rewarding "I don't know". For example, you could give 0.2 points for saying "I don't know", 0 for incorrect, and 1 for correct. This way the model will try to answer but only whe
Re: (Score:2)
Kind of like in democracy. Two uneducated dropouts have more power than one university lecturer.
Yes. The reward system for training humans seems to be similar to the reward system for training AI's in the sense that many (most?) people value an ill informed, confidently delivered opinion, if it confirms their biases, over accurate nuanced, qualified information, whether it comes from a human or a machine!
AI disinformation is poisioning the world (Score:2)
Wow this is very insightful (Score:1)
Re: (Score:2)
OpenAI's models, and most of the LLMs, are trained at least in part by having humans rate their conversations. That's what the "chat" in chatGPT stands for. Humans apparently rate chat partners that make up truthy sounding stuff more highly than chat partners that admit they don't know.
That's a useful finding for a company that makes LLMs. It should be an interesting observation for people who talk to other people too.
Re: (Score:2)
It's the basis of how people get elected and promoted.
Re: (Score:2)
It's not just that. People exhibit overconfidence all the time. From social media and casual conversation to public policy, It's a deep cognitive bias in our species.
Religion is maybe the best example. Don't know what the fuck is going on? Just make up a story that sounds good.
Potential for management positions... (Score:3)
Ability to bullshit one's way up is one of key management skills...
Re: (Score:2)
If AI can replace managers, then not everything is yet lost.
Just like people (Score:2)
Nt
They BS instead of admitting they don't know? (Score:2)
I just had this conversation with my coworker (Score:2)
Precision and Recall - not that fucking hard (Score:2)
How many billions of dollars and thousands of man hours have AI "researchers" cost humanity, just because they evidently forgot to remember a core principle of machine learning?
https://en.wikipedia.org/wiki/... [wikipedia.org]
It's literally how LLMs work (Score:2)
Everything they spit out is a "hallucination." They generate text that is *plausible* in that it echoes patterns that it has encountered in its training. What's really amazing is how frequently the output is *correct.*
Obvious (Score:2)
It reminds me of the ways some students try to talk themselves out of not knowing an answer. Funny part is that I spontaneously use the same techniques with chatgpt to figure out if it know what it is talking about.
GPT is unique though, it's responses are similar to a s
Confabulating (Score:2)
In psychology, this is not hallucination. The proper term for making stuff up like this is "confabulation."
In other words (Score:2)
Just like most humans.
Re: (Score:2)
Enhancement: Just like all politicians.
It works like this (Score:1)
If you optimize for correctness, and reward being wrong equally as much as admitting lack of knowledge - Then the best thing to do is to guess.
If you have a multiple choice question, and you don't know the answer, you will on average have a higher score if you guess rather than if you leave out answering.
If you're a student taking a test, and you don't know the answer, you will try to write whatever you could come up with, that you find is most likely, rather than abstaining from answering. That makes sens