![]() |
|
![]() |
| > is that the model itself doesn't know the difference, and will proclaim bullshit with the same level of confidence
which is a good model for what humans do as well |
![]() |
| So much truth here, very refreshing to see!
About time too, the sooner we can stop the madness the better, building a society on top of this technology is a movie I'd rather not see. |
![]() |
| Maybe they shouldn’t have mixed truthful data with obviously untruthful data in the same training data set?
Why not make a model only from truthful data? Like exclude all fiction for example. |
![]() |
| That wouldn't prevent hallucination. An LLM doesn't know what it doesn't know. It will always try to come up with a response that sounds plausible, based on its knowledge or lack thereof. |
![]() |
| There are at least a couple of examples in the article that you refuse to read that describe hybrids from different families. Sorry, but your purported basic knowledge is wrong. |
![]() |
| If that's true for your use case (it's not true for all) then yes, including LLMs in your design would be a design flaw until they get much much better. |
![]() |
| If you ask about opinions, sure. Because there are no "true" opinions.
If you ask about the capital of France, any answer but "Paris" is objectively wrong, whether given by a human or LLM. |
![]() |
| Paris has not always been the capital of France. Many other cities around France have been capital.
https://en.wikipedia.org/wiki/List_of_capitals_of_France There's practically no subject you could bring up that an LLM wouldn't "hallucinate" or give "wrong" information about given that garbage in -> garbage out, and LLMs are trained on all the garbage (as well as too many facts) they've been able to scrape. The LLM lacks the ability to reason about what century the prompt is asking about, and a guess is all it is programmed to do. Also, if you ask 100 French citizens today what the true capital of France is, you're not always going to get "Paris" as a reply 100 times. |
![]() |
| No, I never said that an LLM would always say "Paris". I said that Paris is the actual correct answer. I don't give LLMs that kind of credit; I'm not sure what I said that made you think that I do. |
![]() |
| Post training includes mechanisms to allow LLMs to understand areas that they should exercise caution in answering. It’s not as simple as you say anymore. |
![]() |
| > but it will never plug that gap
They don't have to be perfect, they just have to be better than humans. And that seems very likely to be achievable eventually. |
![]() |
| > No, if I ask a human about something he doesn't know, the first thing he will think about is not a made up answer, it is "I don't know".
You've just made this up, through. It's not what happens. How would somebody even know that they didn't know without trying to come up with an answer? But maybe more convincingly, people who have brain injuries that cause them to neglect a side (i.e. not see the left or right side of things) often don't realize (without a lot of convincing) the extent to which this is happening. If you ask them to explain their unexplainable behaviors, they'll spontaneously concoct the most convincing explanation that they can. https://en.wikipedia.org/wiki/Hemispatial_neglect https://en.wikipedia.org/wiki/Anosognosia People try to make things make sense. LLMs try to minimize a loss function. |
![]() |
| > A human does not do this.
You obviously had never asked me anything. (Specialy tech questions while drinking a cup of cofee.) If I had a cent for every wrong answer, I'd be already a millionair. |
![]() |
| The US had a president for eight years who was re-elected on his ability to act on his “gut reaction”s.
Not saying this is ideal, just that it isn’t the showstopper you present it as. In fact, when people talk about “human values”, it might be worth reflecting on whether this a thing we’re supposed to be protecting or expunging? "I'm not a textbook player, I'm a gut player.” —President George W. Bush. https://www.heraldtribune.com/story/news/2003/01/12/going-to... |
![]() |
| Humans totally do this if their prefrontal cortex shuts down due to fight or flight response. See eg stage fright or giving bullshit answers in leetcode style interviews. |
![]() |
| Producing text is only the visible end product. The LLM is doing a whole lot behind the scenes, which is conceivably analogous to the thought space from which our own words flow. |
![]() |
| Really out-of-ignorance: Is 'proof' the right word here? A more substantial philosophical counter-argument may be needed, but proof sounds weird in these "metaphysical" (for now) discussions. |
![]() |
| We don’t need to “live with this”. We can just not use them, ignore them, or argue against their proliferation and acceptance, as I will continue doing. |
![]() |
| Technically, you are right. Donald Knuth still doesn't use e-mail, after all.
But for the global "we" entity, it is almost certain that it is not going to heed your call. |
![]() |
| LLMs amplify both intelligence and stupidity, which is why Terence Tao finds them about at the level of a mediocre graduate student (and getting better), and you can’t wait for them to die. |
![]() |
| Disagree - https://arxiv.org/abs/2406.17642
We cover halting problem and intractable problems in the related work. Of course LLMs cannot give answers to intractable problems. I also don’t see why you should call an answer of “I cannot compute that” to a halting problem question a hallucination. |
![]() |
| The idea of hooking LLMs back up to themselves, i.e. giving them token prob information somehow or even giving them control over the settings they use to prompt themselves is AWESOME and I cannot believe that no one has seriously done this yet.
I've done it in some jupyter notebooks and the results are really neat, especially since LLMs can be made with a tiny bit of extra code to generate a context "timer" that they wait before they prompt themselves to respond, creating a proper conversational agent system (i.e. not the walkie talkie systems of today) I wrote a paper that mentioned doing things like this for having LLMs act as AI art directors: https://arxiv.org/abs/2311.03716 |
![]() |
| We can't get rid of hallucinations. Hallucinations are a feature not a bug. A recent study by researchers Jim Waldo and Soline Boussard highlights the risks associated with this limitation. In their analysis, they tested several prominent models, including ChatGPT-3.5, ChatGPT-4, Llama, and Google’s Gemini. The researchers found that while the models performed well on well-known topics with a large body of available data, they often struggled with subjects that had limited or contentious information, resulting in inconsistencies and errors.
This challenge is particularly concerning in fields where accuracy is critical, such as scientific research, politics, or legal matters. For instance, the study noted that LLMs could produce inaccurate citations, misattribute quotes, or provide factually wrong information that might appear convincing but lacks a solid foundation. Such errors can lead to real-world consequences, as seen in cases where professionals have relied on LLM-generated content for tasks like legal research or coding, only to discover later that the information was incorrect. https://www.lycee.ai/blog/llm-hallucinations-report |
![]() |
| If you ask a student to solve a problem while admitting when they don't an answer, they will stop at generating gook for an answer.
LLMs on the other hand regularly spew bogus with high confidence. |
![]() |
| Sounds like a missed STTNG story line. I can imagine that such a “Data,” were we ever to build one, would hallucinate from time to time. |
![]() |
| I don't think it's true that modern LLMs are used to replace human beings in the general case, or that any significant number of people believe they exceed human ability in every relevant factor. |
![]() |
| They did say each token is generated using probability, not certainty, given that there is a chance it produces wrong tokens |
Having a mathematical proof is nice, but honestly this whole misunderstanding could have been avoided if we'd just picked a different name for the concept of "producing false information in the course of generating probabilistic text".
"Hallucination" makes it sound like something is going awry in the normal functioning of the model, which subtly suggests that if we could just identify what went awry we could get rid of the problem and restore normal cognitive function to the LLM. The trouble is that the normal functioning of the model is simply to produce plausible-sounding text.
A "hallucination" is not a malfunction of the model, it's a value judgement we assign to the resulting text. All it says is that the text produced is not fit for purpose. Seen through that lens it's obvious that mitigating hallucinations and creating "alignment" are actually identical problems, and we won't solve one without the other.