Discussion about this post

User's avatar
Garrett MacDonald's avatar

Google sent an email with the subject “Help your children get ready to use Gemini Apps.” In it they said the following:

“Remind your children that Gemini isn’t human. Even though it sometimes talks like one, it can’t think for itself or feel emotions”

I noticed I was surprised that they said Gemini can’t “think for itself.” I think that’s obviously untrue, though I realize there’s a lot of ambiguity in that statement.

Expand full comment
Scott J. Forman's avatar

I think that X post about the Zurich paper misrepresented their findings (for attention points...misaligned!) pretty badly:

> How were they so effective? Before replying, another bot stalked the post history of the target, learned about them and their beliefs, and then crafted responses that would perfectly one-shot them.

That's a more-or-less accurate if overwrought description of the "personalized" condition, but (falsely) implies that this was the *winning* condition. In fact, the "generic" and "personalized" conditions were statistically indistinguishable.

Which is actually kinda interesting...that personalization didn't make any difference at all. Not to say that it can't or won't in the future. But in this paper, nope.

Also interesting that their fine-tuned 4o significantly underperformed. But hard to tell whether that's because the fine-tuning didn't help or because they didn't run it through the same ensemble scaffolding, for unclear reasons.

Expand full comment
38 more comments...

No posts