Hacker News new | past | comments | ask | show | jobs | submit login

A statistical model which is instructed to output the token that is most likely to come next doesn’t have “confidence” in its choice based on the distribution of possible tokens. We might, but it cannot. A statistical model cannot be confident or unsure. It has no mind.

It also has no concept of what it means for the choice of token to be an “error” or not, or what a “correct” answer would be.




The model does not "output the token that is most likely to come next". The model provides a list of probabilities and the sampler algorithm picks one; those are two different components.


The point is that neither the model nor the sampler algorithm can possibly have “confidence” in its behaviour or the system’s collective behaviour.

If I put a weight on one side of a die, and I roll it, the die is not more confident that it will land on that side than it would be otherwise, because dice do not have the ability to be confident. Asserting otherwise shows a fundamental misunderstanding of what a die is.

The same is true for LLMs.


I think it's better to say that it's not grounded in anything. (Of course, the sampler is free to verify it with some external verifier, and then it would be.)

But there are algorithms with stopping conditions (Newton-Raphson, gradient descent), and you could say that an answer is "uncertain" if it hasn't run long enough to come up with a good enough answer yet.


If we run the Newton-Raphson algorithm on some input and it hasn’t run long enough to come up with a good enough answer yet, then we are uncertain about the answer. It is not the case that the algorithm is uncertain about the answer. It would make no sense to make any claims about the algorithm’s level of certainty, because an algorithm does not have the capacity to be certain.


I'm not the one doing the arithmetic here, I've outsourced it to the computer. So I don't have any calculated uncertainty because I'm not paying enough attention to know how much progress it's made.


The important part is that the algorithm doesn’t either.


"confidence" doesn't have to be an emotional state. It's essentially just another word for "probability" here - any model's confidence of X is the probability it yields for X. Isn't this common terminology?


It may be terminology that some people use in that way, but it’s becoming increasingly common for people describing LLMs to use such terminology to mean that the LLM literally has the capacity for understanding.

Personally, until recently I can only recall people saying things along the lines of “applying the model indicates that we can state this fact about the data with this much confidence”, never “the model has this much confidence” in some truth statement, especially one independent of its training data.


All the research we have on this points pretty blatantly to everything you've just said being untrue.

Yes, LLMs have a pretty good idea of the uncertainty and truth of their predictions internally. https://news.ycombinator.com/item?id=41418486


You’re missing my point. Take one of the articles described in that comment, titled “The Internal State of an LLM Knows When It's Lying”. It states “In this paper, we provide evidence that the LLM's internal state can be used to reveal the truthfulness of statements.” Both of these are untrue, for a number of reasons.

- An LLM knowing when it is lying is not the same thing as its internal state being able to “reveal the truthfulness of statements”. The LLM does not know when it is lying, because LLMs do not know things.

- It is incapable of lying, because lying requires possessing intent to lie. Stating untrue things is not the same as lying.

- As the paper states shortly afterwards, what it actually shows is “given a set of test sentences, of which half are true and half false, our trained classifier achieves an average of 71% to 83% accuracy”. That’s not the same thing as it being able to “reveal the truthfulness of statements”.

No intellectually honest person would claim that this finding means an LLM “knows when it is lying”.


I'm not missing your point. I just don't think you're making one.

You keep saying the same nonsense over and over again. A LLM does not know things so... What kind of argument is that ? You're working backwards from a conclusion that is nothing but your own erroneous convictions on what a "statistical model" is and are undertaking a whole lot of mental gymnastics to stay there.

There are a lot of papers there that all try to approach this in different ways. You should read them and try to make an honest argument and that doesn't involve "This doesn't count because - claim that is in no way empirically or theoretically validated."


You are the one claiming that LLMs are conscious, so it falls to you to prove it.

I argued that LLMs do not have the capacity to have ideas or to know things, and you tried to prove me wrong by providing examples of papers that show, for example, that LLMs have internal states that can be used to predict the likelihood that what they will output will be facts. But that doesn’t disprove what I said, because that’s not what it means to have ideas or know things. By definition, only conscious beings can do those things.


>You are the one claiming that LLMs are conscious, so it falls to you to prove it.

If a machine is doing things previously before ascribed to "conscious beings" then it's on you to tell me why the machine is not conscious. Hopefully something other than the circular - "It cannot be conscious so it is not conscious".

But whatever. I hadn't quite realized this had devolved into a debate on consciousness. I think that's on me but I have no interest in a back and forth on such an ill-defined, ill-understood concept.

You don't know what consciousness is, what is required of it or what makes it tick in you, you have no way of proving one way or another anybody else has it. It's extremely silly then don't you think to make such bold declarations on what doesn't have it ? especially with circular arguments.

What difference does it make if you won't call it conscious if it does anything a conscious being does ? That's just semantics.


You’re still failing to understand that a model being able to output a prediction of something is not the same thing as it “knowing” that thing. The Newton-Raphson method doesn’t “know” what the root of a function is, it just outputs an approximation of it.

> It’s extremely silly then don’t you think to make such bold declarations on what doesn’t have it?

I don’t find it particularly bold to respond to your assertion that a piece of mathematics is sentient life by stating that you haven’t proven that it is, and that in the absence of that proof, the most rational position is to continue to believe that it is not, as we have done for millennia. The burden of proof is on you.

> if it does anything a conscious being does

You haven’t shown that it can do anything that only conscious beings can do.

Being able to generate a passable approximation of text that might follow some prompt doesn’t mean that it understands the prompt, or its answer. As an obvious example, if you give LLMs maths problems, they change their answers if you change the names of the people in the question. They’re not actually doing maths.

> Notice anything? It’s not just that the performance on MathGLM steadily declines as the problems gets bigger, with the discrepancy between it and a calculator steadily increasing, it’s that the LLM based system is generalizing by similarity, doing better on cases that are in or near the training set, never, ever getting to a complete, abstract, reliable representation of what multiplication is.[0]

[0] https://garymarcus.substack.com/p/math-is-hard-if-you-are-an...


>You’re still failing to understand that a model being able to output a prediction of something is not the same thing as it “knowing” that thing. The Newton-Raphson method doesn’t “know” what the root of a function is, it just outputs an approximation of it.

That is your assertion. I'm not failing to understand anything. I'm simply telling you that you are stating an unproven assertion. This is why i don't like to debate consciousness.

Unless you believe in magic then the only thing that would stop whatever is running 'Newton-Ralph' from "knowing" roots if you are even right is that's it's not the kind of computation that "knows", not because it's a computation.

>I don’t find it particularly bold to respond to your assertion that a piece of mathematics is sentient life by stating that you haven’t proven that it is, and that in the absence of that proof, the most rational position is to continue to believe that it is not, as we have done for millennia. The burden of proof is on you.

The brain computes and unless you believe in a soul or something similar then that is all the brain does to produce consciousness. Computation is substrate independent[0]. Whether it is chemical reactions and nerve impulses or transistors in chips or even pulleys, it does not at all matter what is performing this computation.

Consciousness is clearly an emergent property. Your neurons are not conscious and they do not do conscious things and yet you believe you are conscious. "piece of mathematics" is entirely irrelevant here.

>You haven’t shown that it can do anything that only conscious beings can do. Being able to generate a passable approximation of text that might follow some prompt doesn’t mean that it understands the prompt, or its answer.

I know LLMs understand because of the kind of responses i get to the kind of queries i give them. This is how we probe and test understanding in humans.

>As an obvious example, if you give LLMs maths problems, they change their answers if you change the names of the people in the question.

No they don't. If you'd actually read that apple paper (i assume that's what's you are referring to), you would see that GPT-4o, o1-mini and o1-prievew do not shift above or below the margin of error numbers on 4/5 on the synthetic benchmarks they created. Definitely not for the ones that were just changing of names. So this is blatantly wrong. Changing names literally does nothing for today's state of the art LLMs

That Gary Marcus blog is idiotic but i don't expect much from gary marcus. There is not a single human on this planet that can perform arithmetic unaided (no calculator/writing down numbers) better than SOTA LLMs today. I guess humans don't understand or do math.

Not to mention that you can in fact train transformers that will generalize perfectly on addition.[1]

[0] https://www.edge.org/response-detail/27126

[1]https://www.alignmentforum.org/posts/N6WM6hs7RQMKDhYjB/a-mec...




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: