A more suitable term would be confabulation, which is what humans do when due to a memory error (eg. due to Korsakoff syndrome) we produce distorted memories of oneself. This may sometimes sound very believable to outsiders; the comparison with LLMs making stuff up is rather apt!
So please call it confabulating instead of hallucinating when LLMs make stuff up.
The fact I had to look it up to make sure, and that it isn't the primary definition, makes this a bad alternative to a well understood word which has already become established.
If you do want to use another term for laypeople, I think "bullshitting" or "BSing" would have connotations that are more relevant than "hallucinating".
That's not to say they're not useful. The ability to BS is well regarded among humans as long as you, as a consumer, have a decent BS detector. And like a good BS artist, if you stay within their area of expertise they can be really useful. Its when you ask them something that they should or almost know that they start to be full of s**.
> LLMs have no sensors to experience the world with (other than their text input) and (probably) don’t even have a subjective experience in the same way humans do.
Computers are not subjects, period, and to attribute intentionality to them is nothing short of projection and superstition. LLMs do not somehow transcend what a computer is. Computers are machines that transform conventional representations into other conventional representations, but these representations require human interpretation to have meaning (during which the representation is related to the conventionally assigned meaning which is itself the terminus that requires no further interpretation). That meaning or content is the intentionality in question. Your won’t find it in the machine, by definition.
A source of popular misunderstanding probably comes from the confusion of what the computer is objectively doing with the metaphorical language being used to talk about it. Even “confabulation” is, strictly speaking, metaphorical or analogical, but within the analogical schema we’re using, it is a better fit than hallucination.
Besides both the general public and academics have been calling it “hallucination” for well over a year now. I’m sorry to say this ship has long sailed.
Couldn't llms (or rather generative transformers) sort of confabulate their own subjective experience? If you added other senses/input, just text, that could all act on neurons, a 'memory', a feeling of passing time (internal clock or whatever, this is weirdly a sense every viable human get, as do all complex animal lifeforms). Is this a technical possibility? If a GPT was able to remember and 'internalize' recent input (I mean it passed recent input in its training function, not at all like ChatGPT 'remember' previous input/output that he experienced very recently) and it's output, and you asked it 'why was that your response', would it confabulate too? Would it trick itself? And could it experience 'déjà vu' if the 'internalization' bugged?
Why? LLMs don't have memories any more than they have sensors.
"ChatGPT: these are not hallucinations – they’re fabrications and falsifications" https://www.nature.com/articles/s41537-023-00379-4#author-in...
"Chatbot Confabulations Are Not Hallucinations" https://jamanetwork.com/journals/jamainternalmedicine/articl...
While I agree in principle, I'm not sure 'confabulating' is doing enough work. My hypothesis is given two words--one more accurate and one less accurate to describe the same phenomena--people will choose the more expansive or imaginative notion. That is, people will choose a word which generates more ideas.
We are in a time of expanding imagination about what LLMs and "AI" will do in the future. By definition 'hallucinating' LLMs are ridiculous, but the thoughts extending from 'hallucinating' label are nevertheless expansive in a similar way to our general overall feelings towards these technologies. Whereas 'confabulation' does nothing for my imagination.
Say for instance I summarize something and want to check that the result doesn't contain hallucinations (confabulations :)) or more specifically that the summary contains only information present in original text. What's current state of the art for something like this and how well does it perform? I've read some about entailment models and fine tuned LLMs for this sort of thing but haven't found many great resources.
The thing is, “confabulating” will never stick. People like myself will enter discussions about it and insist that it won’t stick, and because of that, I certainly won’t start using that term and hopefully I’ll convince a few others not to either.
* this is the prevailing secular trend, at least in the US
Tldr; language changes over time.
The idea of “precise language, and that someone is “using words wrong” is not correct, and history shows, repeatedly, obsessively trying to enforce “correctness” in language doesn’t work.
So… this is no different from telling people not to use “like, someone told me…” in sentences.
It won’t stick, and perusing it is probably meaningless.
If people call it hallucinations; that’s what they are.
That’s what language is.
There’s some deep irony in talking about this with language models.
seriously though, hallucinating makes sense to me because it genuinely feels like it's seeing things that doesn't exist.
for example, it comes up with non-existent postgresql functions.
that's hallucinations right there, i sometimes wonder what gpt-3.5 is smoking - i wanna try it.