Discussion about this post

User's avatar
Michael Sullivan's avatar

I have considered open-sourcing a couple of small codebases that I vibe coded, but then decided that was a category error: the entire point of those projects is that they are personal tools. If someone wants a similar tool, instead of downloading mine from Github, they should get AI to code their own, tuned to their tastes. That's what's special about these things: they aren't particularly good or polished products, they're ones designed to your personal wants and needs.

So that's my little contribution to a lack of explosion of new Github repos.

Mike's avatar

Could LLM agency be held back by chatbot/instruct training?

I've been wondering this for a while and have a couple reasons to suspect so but the short version is that the models are always attempting a 1:1 match of user message to instruction-following reply, and considering "everything affects everything", I expect this instills some literal servitude in the models, along with a bunch of other weirdness.

The "thinking model" is an interesting hack, but 'agency' seems to mostly come from tool use, (extensive) prompting and a while loop.

So, is there a different way, starting from a base model, to actually get to an agent model, where thinking and goal-following is the core behaviour and "chatting to a user" is just another tool call?

It seems obvious enough, but the lack of models like this (there could be small ones I don't know about?) suggests that either it doesn't work (well), or alignment/safety is more difficult, or the labs are just keeping it cooking internally. Or path dependence.

17 more comments...

No posts

Ready for more?