Benchmarks used to rank AI models are several years old, often sourced from amateur websites, and, experts worry, lending automated systems a dubious sense of authority
I’m not saying humans and LLMs generate language the same way.
I’m not saying humans and LLMs don’t generate language the same way.
I’m saying I don’t know and I haven’t seen clear data/evidence/papers/science to lean one way or the other.
A lot of people seem to believe humans and LLMs don’t generate language the same way. I’m challenging that belief in the absence of data/evidence/papers/science.
You’re actually incorrect in regards to Russell’s teapot in this instance. The correct approach is to admit to yourself and others you don’t know. Not to assume a negative became you can’t prove a positive, if you can’t prove the negative either.
I know I don’t know, but this is a continuous system and the probability of something being in one particular state is infinitely small ; the probability of it being in certain range of that particular state is, ahem, not, but with the amount of moving things in LLMs and in human brains there are most likely quite a few radical differences between laws describing them.
Why am I incorrect? You can’t disprove that there isn’t that teapot flying at a certain orbit as well. Or you can, but not for all such statements.
What would be the criterion for saying that yes, human brain works with language just in the same way as LLMs do? What would be “same”? Logic exists inside defined constraints in the continuous world.
Unless you define what would prove something, you can’t disprove it, but it’s also not a scientific hypothesis. That’s Popper’s criterion.
I’m not saying humans and LLMs generate language the same way.
I’m not saying humans and LLMs don’t generate language the same way.
I’m saying I don’t know and I haven’t seen clear data/evidence/papers/science to lean one way or the other.
A lot of people seem to believe humans and LLMs don’t generate language the same way. I’m challenging that belief in the absence of data/evidence/papers/science.
Like going out and meeting a dino - 50% yes, 50% no. It’s a joke.
Russell’s teapot again.
You’re actually incorrect in regards to Russell’s teapot in this instance. The correct approach is to admit to yourself and others you don’t know. Not to assume a negative became you can’t prove a positive, if you can’t prove the negative either.
I know I don’t know, but this is a continuous system and the probability of something being in one particular state is infinitely small ; the probability of it being in certain range of that particular state is, ahem, not, but with the amount of moving things in LLMs and in human brains there are most likely quite a few radical differences between laws describing them.
Why am I incorrect? You can’t disprove that there isn’t that teapot flying at a certain orbit as well. Or you can, but not for all such statements.
What would be the criterion for saying that yes, human brain works with language just in the same way as LLMs do? What would be “same”? Logic exists inside defined constraints in the continuous world.
Unless you define what would prove something, you can’t disprove it, but it’s also not a scientific hypothesis. That’s Popper’s criterion.