Discussion about this post

User's avatar
Stef's avatar

Hype, yes, and "...its just Claude with N tools and X and Y..." and okay, but if that's all it takes to run our lives, then, yeah, we're pretty close to being cooked.

We're not all that complicated really. Most people really *could* be run inside a Very Small Shell Script.

Or as Anthony Hopkins in "Westworld" put it:

"Intelligence, it seems, is just like a Peacock really..."

https://youtu.be/OjsGJOXhPGU

Expand full comment
Erik's avatar

"Yes, in two years AI agents are going to be absurdly powerful. Wait for it."

With every release of an agent that doesn't really work I update a little against this, actually. Not that I expected them to Really Work yet - but I'm kind of convinced by Colin Fraser's point that you can't really get goal-oriented behaviors out of LLMs yet: if they have goals at all, it's to simulate text, and simulating text that looks like solving problems doesn't suffice to solve lots of hard problems. It won't be critical enough of its own work, it won't have high enough standards.

Specifically I update against something like "a 10x scaled up GPT-4.5 with CoT and great scaffolding with aggressive agent-flavored RL" could produce good-enough agents. In that world, the gap probably requires a transformers-sized breakthrough, so to speak (though given the manpower thrown at this problem and the crazy-high benefits we might get that).

Expand full comment
7 more comments...

No posts