Discussion about this post

User's avatar
Richard Demsyn-Jones's avatar

I see what you have, yet I've ended up very bullish on current and future Ai capabilities. I'll lay out why. I've been impressed with how capable LLMs are despite their lack of recurrence. Whether much of it qualifies as insight or not is debatable, but we could have the same critique of human writing too.

There were a few factors that I didn't anticipate or didn't anticipate to this degree:

1. The size, in both depth and breadth, of leading LLMs allows them to encode a lot of logic that substitutes for loops. With every token in the context window interacting with each other, and that happening throughout a large number of layers, LLMs can subsume a lot of reasoning skills and even have pieces of logic that are like bounded unwound loops. Those loop-like constructs can't be truly infinite, but most human reasoning—even with the tool of writing—likely doesn't loop too many times either.

2. Increased context window size allowing for LLMs to consider and interact more information. Context windows have grown immensely without a massive trade-off in the models' ability to use their context thoroughly.

3. The invention of deep reasoning models. Reasoning models learn a hack that they can essentially reset and reconsider, starting fresh over their context window that includes their input (or nearly as much of it, if it was very long) plus newly constructed output that the models generated. This is basically a loop added around the process combined with optimizing the models for using that loop capability.

Those all interact with each other. I don't see an inherent bound on how insightful LLMs can become. Those techniques are powerful, and they make me optimistic that people will develop more innovations that will also strengthen AI capabilities.

Expand full comment
1 more comment...

No posts