Introspection in LLMs: A Proposal For How To Think About It, And Test For It
lesswrong.com·17h
Flag this post

Published on November 22, 2025 2:52 PM GMT

The faculty of introspection in LLMs is an important, fascinating, increasingly popular, and somewhat underspecified object of study. Studies could benefit from researchers being explicit about what their assumptions and definitions are; the goal of this piece is to articulate mine as clearly as possible, and ideally stimulate further discussion in the field.

~~~~~~~~~~~~~~~~~~~~~~~~~~~~

My first assumption is that introspection means the same thing whether we are talking about humans, animals, or AI, although it may be implemented differently; this enables us to take prior conceptual and behavioral work on human and animal introspection and apply it to LLMs. The <a href=“https://plato.stanford.edu/entries/introspectio…

Similar Posts

Loading similar posts...