Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

"Hallucinations" implies that someone isn't of sound mental state. We can argue forever about what that means for a LLM and whether that's appropriate, but I think it's absolutely the right attitude and approach to be taking toward these things.

They simply do not behave like humans of sound minds, and "hallucinations" conveys that in a way that "confabulations" or even "bullshit" does not. (Though "bullshit" isn't bad either.)



I disagree with this take because LLMs are, always, hallucinating. When they get things right it’s because they are lucky. Yes, yes, it’s more complicated than that, but the essence of LLMs is that they are very good at being lucky. So good that they will often give you better results than random search engine clicks, but not good enough to be useful for anything important.

I think what calling the times they get things wrong hallucinations is largely an advertising trick. So that they can sort of fit the LLMs into how all IT is sometimes “wonky” and sell their fundamentally flawed technology more easily. I also think it works extremely well.


But the point is, isn't hallucinating about having malformed, altered or out of touch input rather than producing inaccurate output yourself?

It is the memory pathways leading them astray. It could be thought of a memory system that at certain point any longer can't be fully sure if whatever connections they have are from actually being trained or it or created accidentally.


> isn't hallucinating about having malformed, altered or out of touch input rather than producing inaccurate output yourself?

I suppose so, in the sense that someone could simply be lying about pink elephants instead of seeing them. However it's hard to argue that the machine knows the "right" answer and is (intelligently?) deceiving us.

> It is the memory pathways leading them astray.

I don't think it's a "memory" issue as much as a "they don't operate the way we like to think they do" issue.

Suppose a human is asked to describe different paintings on the wall of an art gallery. Sometimes their statements appear valid and you nod along, and sometimes the statements are so wrong that it alarms you, because "this person is hallucinating."

Now consider how the entire situation is flipped by finding out one additional fact... They're actually totally blind.

Is it a lie? Is it a hallucination? Does it matter? Either way you must dramatically re-evaluate what their "good" outputs really mean and whether they can be used.


To me it's more like, imagine that you have read a lot of books throughout your life, but then someone comes in and asks a question from you and you try to answer from memory, but you get beaten when you say something like "I don't know", and you get rewarded if you answer accurately. You do get beaten if you answer inaccurately, but eventually you learn that if you just say something, you might just be accurate and you will not get beaten. So you just always learn to answer to the best of your knowledge, while never saying that you specifically don't know, because it decreases chances of getting beat up. You are not intentionally lying, you are just hoping that whatever you say is accurate to the best you can do according to the neural connections you've built up in your brain.

Like you ask me for a birthdate of some obscure political figure from history? I'm going to try to feel out what period in history the name might feel like to me and just make my best guess based on that, then say some random year and a birthdate. It just has the lowest odds of being beaten. Was I hallucinating? No, I was just trying to not get beaten.


> When they get things right it’s because they are lucky.

This is transparently wrong. It gets so many things right in a response that the few things it gets wrong are tremendously frustrating. I think people underestimate how much correct "knowledge about the world" is expressed in a typical chat gpt response and focus only on the parts that are incorrect.

If it were wrong about _everything_ at rates no better than chance, we wouldn't even be having this conversation because nobody would be using them.


To offer a satirical analogy: "Lastly, I want to reassure investors and members of the press that we take these concerns very seriously: Hindenburg 2 will only contain only normal and unreactive hydrogen gas, and not the rare and unusual explosive kind, which is merely a temporary hurdle in this highly dynamic and growing field."

Edit: It retrospect, perhaps a better analogy would involve gasoline, as its explosive nature is what's being actively being exploited in normal use.


Yes (to the edit), an analogy with making planes safer by only using non-flammable fuels is perfect.


I expect most people have already filled in the blanks, but for completeness:

"Lastly, I want to reassure investors and members of the press that we take these concerns very seriously: The Ford Pinto-II will only contain only normal and stable gasoline, and not the rare and unusual burning kind, which is merely a temporary hurdle in this highly dynamic and explos--er--fast growing field."


I don't really immediately link "Hallucinations" with "Unsound mind" - most people I know have experienced auditory hallucinations - often things like not sure if the doorbell went off, or if someone said their name.

And I couldn't find a single one of my friends who hadn't experienced "phantom vibration syndrome".

Both I'd say are "Hallucinations", without any real negative connotation.


Hallucinations implies that they do behave like a human mind. Why else would you use the word if you were not trying to draw this parallel?


We're stuck with metaphors for human behavior because the way LLMs operate is so alien and counterintuitive, yet similar enough to human behavior, that we haven't yet developed suitable language to describe it. "Hallucination" gets the point across in general terms, at least.


"Sound" minds for humans is graded on a curve, and this trick is not acknowledged, or popular.


How about "dream-reality confusion (DRC)" ?


There is no dream-reality separation in an LLM, or really any conception of dreams or reality, so I don't think the term makes sense. Hallucination works fine to describe the phenomenon. LLMs work by coalescing textual information. LLM hallucinations occur due to faulty or inappropriate coalescence of information, which is similar to what occurs with actual hallucinations.


"Incoherence" seems like a far more natural fit for what you're describing than a human with a sensory delusion or psychosis.


Bullshit is the most descriptive one.

LLMs don't do it because they are out of their right mind. They do it because every single answer they say is invented caring only about form, and not correctness.

But yeah, that ship has already sailed.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: