self-assessment

  • LLMs Reading Their Own Reasoning


    We need an LLM that can read it's own thoughts.Many large language models (LLMs) that claim to have reasoning capabilities cannot actually read their own reasoning processes, as indicated by the inability to interpret tags in their outputs. Even when settings are adjusted to show raw LLM output, models like Qwen3 and SmolLM3 fail to recognize these tags, leaving the reasoning invisible to the LLM itself. However, Claude, a different LLM, demonstrates a unique ability to perform hybrid reasoning by using tags, allowing it to read and interpret its reasoning both in current and future responses. This capability highlights the need for more LLMs that can self-assess and utilize their reasoning processes effectively, enhancing their utility and accuracy in complex tasks.

    Read Full Article: LLMs Reading Their Own Reasoning