Discussion about this post

User's avatar
Arbituram's avatar

Worth highlighting good practice here:

"As part of our continued effort to partner with external experts, joint pre-deployment testing of the new Claude 3.5 Sonnet model was conducted by the US AI Safety Institute (US AISI) and the UK Safety Institute (UK AISI)."

https://www.anthropic.com/news/3-5-models-and-computer-use

Expand full comment
Rai Sur's avatar

I thought your position was that since no one is putting AI in boxes at all (in fact they're maximally connecting them to the internet), reducing overhang by doing unhobbling like Claude's computer use is actually kinda desirable?

Imagine a world where networked base-models are getting more powerful but not much effort is put into squeezing all their utility out before moving onto training a more powerful model. This seems pretty dangerous because when people do put effort into unhobbling a later model, it may leapfrog all sorts of capabilities at once.

Expand full comment
7 more comments...

No posts