zlacker

[parent] [thread] 12 comments
1. Volund+(OP)[view] [source] 2025-05-06 16:59:08
> Many attempts at making them refuse to answer what they don't know caused them to refuse to answer things they did in fact know.

Are we sure they know these things as opposed to being able to consistently guess correctly? With LLMs I'm not sure we even have a clear definition of what it means for it to "know" something.

replies(2): >>redox9+s2 >>ajross+j7
2. redox9+s2[view] [source] 2025-05-06 17:13:44
>>Volund+(OP)
Yes. You could ask for factual information like "Tallest building in X place" and first it would answer it did not know. After pressuring it, it would answer with the correct building and height.

But also things where guessing was desirable. For example with a riddle it would tell you it did not know or there wasn't enough information. After pressuring it to answer anyway it would correctly solve the riddle.

The official llama 2 finetune was pretty bad with this stuff.

replies(1): >>Volund+nI
3. ajross+j7[view] [source] 2025-05-06 17:42:36
>>Volund+(OP)
> Are we sure they know these things as opposed to being able to consistently guess correctly?

What is the practical difference you're imagining between "consistently correct guess" and "knowledge"?

LLMs aren't databases. We have databases. LLMs are probabilistic inference engines. All they do is guess, essentially. The discussion here is about how to get the guess to "check itself" with a firmer idea of "truth". And it turns out that's hard because it requires that the guessing engine know that something needs to be checked in the first place.

replies(2): >>myname+68 >>Volund+eF
◧◩
4. myname+68[view] [source] [discussion] 2025-05-06 17:46:49
>>ajross+j7
Simple, and even simpler from your own example.

Knowledge has an objective correctness. We know that there is a "right" and "wrong" answer and we know what a "right" answer is. "Consistently correct guesses", based on the name itself, is not reliable enough to actually be trusted. There's absolutely no guarantee that the next "consistently correct guess" is knowledge or a hallucination.

replies(2): >>ajross+m9 >>fwip+oe
◧◩◪
5. ajross+m9[view] [source] [discussion] 2025-05-06 17:55:14
>>myname+68
This is a circular semantic argument. You're saying knowledge is knowledge because it's correct, where guessing is guessing because it's a guess. But "is it correct?" is precisely the question you're asking the poor LLM to answer in the first place. It's not helpful to just demand a computation device work the way you want, you need to actually make it work.

Also, too, there are whole subfields of philosophy that make your statement here kinda laughably naive. Suffice it to say that, no, knowledge as rigorously understood does not have "an objective correctness".

replies(2): >>myname+cd >>Volund+qG
◧◩◪◨
6. myname+cd[view] [source] [discussion] 2025-05-06 18:19:13
>>ajross+m9
I mean, it clearly does based on your comments showing a need for a correctness check to disambiguate between made up "hallucinations" and actual "knowledge" (together, a "consistently correct guess").

The fact that you are humanizing an LLM is honestly just plain weird. It does not have feelings. It doesn't care that it has to answer "is it correct?" and saying poor LLM is just trying to tug on heartstrings to make your point.

replies(1): >>ajross+cj
◧◩◪
7. fwip+oe[view] [source] [discussion] 2025-05-06 18:28:00
>>myname+68
So, if that were so, then an LLM possess no knowledge whatsoever, and cannot ever be trusted. Is that the line of thought you are drawing?
◧◩◪◨⬒
8. ajross+cj[view] [source] [discussion] 2025-05-06 18:56:12
>>myname+cd
FWIW "asking the poor <system> to do <requirement>" is an extremely common idiom. It's used as a metaphor for an inappropriate or unachievable design requirement. Nothing to do with LLMs. I work on microcontrollers for a living.
◧◩
9. Volund+eF[view] [source] [discussion] 2025-05-06 21:27:03
>>ajross+j7
> What is the practical difference you're imagining between "consistently correct guess" and "knowledge"?

Knowing it's correct. You've just instructed it not to guess remember? With practice people can get really good at guessing all sorts of things.

I think people have a serious misunderstanding about how these things work. They don't have their training set sitting around for reference. They are usually guessing. Most of the time with enough consistency that it seems like they "know'. Then when they get it wrong we call it "hallucinations". But instructing then not to guess means suddenly they can't answer much. There no guessing vs not with an LLM, it's all the same statistical process, the difference is just if it gives the right answer or not.

replies(2): >>Subicu+C11 >>ajross+N31
◧◩◪◨
10. Volund+qG[view] [source] [discussion] 2025-05-06 21:36:30
>>ajross+m9
> You're saying knowledge is knowledge because it's correct, where guessing is guessing because it's a guess.

Knowledge is knowledge because the knower knows it to be correct. I know I'm typing this into my phone, because it's right here in my hand. I'm guessing you typed your reply into some electronic device. I'm guessing this is true for all your comments. Am I 100% accurate? You'll have to answer that for me. I don't know it to be true, it's a highly informed guess.

Being wrong sometimes is not what makes a guess a guess. It's the different between pulling something from your memory banks, be they biological or mechanical, vs inferring it from some combination of your knowledge (what's in those memory banks), statistics, intuition, and whatever other fairy dust you sprinkle on.

◧◩
11. Volund+nI[view] [source] [discussion] 2025-05-06 21:52:05
>>redox9+s2
> After pressuring it, it would answer with the correct building and height.

And if you bully it enough on something nonsensical it'll give you a wrong answer.

You press it, and it takes a guess even though you told it not to, and gets it right, then you go "see it knew!". There's no database hanging out in ChatGPT/Claude/Gemini's weights with a list of cities and the tallest buildings. There's a whole bunch of opaque stats derived from the content it's been trained on that means that most of the time it'll come up with the same guess. But there's no difference in process between that highly consistent response to you asking the tallest building in New York and the one where it hallucinates a Python method that doesn't exist, or suggests glue to keep the cheese on your pizza. It's all the same process to the LLM.

◧◩◪
12. Subicu+C11[view] [source] [discussion] 2025-05-07 01:03:46
>>Volund+eF
Maybe LLM's know so much that it makes it difficult to feel the absence. When someone asks me about the history of the Ethiopian region, I can at most recall very few pieces of information, and critically, there is an absence of feelings of familiarity. In memory research, familiarity signal is can prompt continued retrieval attempts, but importantly absence of that signal can suggest that further retrieval attempts would be fruitless, and that you know that you don't know. Maybe knowing so much means that there is near saturation for stop tokens...or that llms need to produce a familiarity like scoring of the key response of interest.
◧◩◪
13. ajross+N31[view] [source] [discussion] 2025-05-07 01:27:42
>>Volund+eF
> Knowing it's correct.

I love the convergence with philosophy here.

This is the second reply that's naively just asserted a tautology. You can't define "knowledge" in terms of "knowing" in the sense of the English words; they're the same word! (You can, I guess, if you're willing to write a thesis introducing the specifics of your jargon.)

In point of fact LLMs "know" that they're right, because if they didn't "know" that they wouldn't have told you what they know. Which, we all agree, they do know, right? They give answers that are correct. Usually.

Except when they're wrong. But that's the thing: define "when they're wrong" in a way rigorous enough to permit an engineering solution. But you really can't, for the same reason that you can't prevent us yahoos on the internet from being wrong all the time too.

[go to top]