Reasoning models don't so much think as navigate
The new language models are children of the reasoning revolution, and they stream out these long, circuitous thinking traces. They are said to be applying more compute to our questions and challenges.
This is subtle, but that “more” isn’t particularly about thinking harder. Rather, it’s about thinking in the right direction. It is not the gas pedal, but the steering wheel —
The reasoning revolution depends, in part, on the unreasonable effectiveness of specific words: twists like “but wait” and “actually”, which operate as powerfully as magic spells. (The English department NEEDS to get into the game with this stuff.) Is the phrase “but wait” really a white-hot kernel of intellectual effort? No. It’s a sign planted in the ground, pointing THAT-A-WAY, towards a particular kind of document that humans find useful.
(Don’t mistake precision for minimization. I’m not dismissively saying, these are just documents; I am plainly observing, these are documents. If you don’t think documents are cool, even sometimes cosmic, that’s on you!)
Notice that, as in real life, directions aren’t always correct. It’s likely that you have by now watched a language model walk in circles, “but wait”-ing itself back around, and around, and around again …
Recent research from Apple talks about “forks” in the road, with “distractors” that can lead a model in the wrong direction.
Here’s more evidence for the navigation argument: base models can already do the things reasoning models can do … it just takes them much longer to arrive in the correct regions of high-dimensional space. Base models are fine thinkers, but cruddy navigators.
The single forward pass of a language model runs on its own, refracting a context window into an array of probabilities; that’s all “the model” ever does. However, each forward pass can “stand on the shoulders of giants”, taking direction from previous passes, bringing its brief labors into better alignment with the desires of the human operator, way out there.
As usual, observations about language models raise questions about human minds. Do we think harder mostly by thinking in the right direction? I think the answer is sometimes yes —
(This post is related to the latest edition of my pop-up newsletter on AI.)
To the blog home page