Anthropic’s latest research shows the chain of thought reasoning shown isn’t trustworthy anyway. It’s for our benefit, and doesn’t match the actual reasoning used internally 1:1.
As you say, hallucinating can be solved by adding meta-awareness. Seems likely to me we’ll be able to patch the problem eventually. We’re just starting to understand why these models hallucinate in the first place.
Anthropic’s latest research shows the chain of thought reasoning shown isn’t trustworthy anyway. It’s for our benefit, and doesn’t match the actual reasoning used internally 1:1.
deleted by creator
As you say, hallucinating can be solved by adding meta-awareness. Seems likely to me we’ll be able to patch the problem eventually. We’re just starting to understand why these models hallucinate in the first place.
deleted by creator