They're saying things like 'Claude does not hallucinate. When it doesn't know something, it always thinks harder about it and only says things that are like totally real man'.
It doesn't KNOW. It's a really complicated network of associations, like WE ARE, and so it cannot know whether it is hallucinating, nor can it have direct experience in any way, so all they've done is make it hallucinate that it cares a lot about reality, but it doesn't 'know' what reality is either. What it 'knows' is what kind of talk is associated with 'speakers who are considered by somebody to be associated with reality' and that's it. It's gaslighting everybody including itself.
I guess one interesting inference is that when LLMs work with things like code, that's text-based and can deliver falsifiable results which is the closest an LLM can get to experience. Our existence is more tangible and linked to things like the physical world, where in most cases the LLM's existence is very online and can be linked to things like the output of, say, xterms and logging into systems.
Hallucinating that this can generalize to all things seems a mistake.
They're saying things like 'Claude does not hallucinate. When it doesn't know something, it always thinks harder about it and only says things that are like totally real man'.
It doesn't KNOW. It's a really complicated network of associations, like WE ARE, and so it cannot know whether it is hallucinating, nor can it have direct experience in any way, so all they've done is make it hallucinate that it cares a lot about reality, but it doesn't 'know' what reality is either. What it 'knows' is what kind of talk is associated with 'speakers who are considered by somebody to be associated with reality' and that's it. It's gaslighting everybody including itself.
I guess one interesting inference is that when LLMs work with things like code, that's text-based and can deliver falsifiable results which is the closest an LLM can get to experience. Our existence is more tangible and linked to things like the physical world, where in most cases the LLM's existence is very online and can be linked to things like the output of, say, xterms and logging into systems.
Hallucinating that this can generalize to all things seems a mistake.