Discussion about this post

User's avatar
Dave Orr's avatar

Not to nitpick language, but "prompt injection" means a malicious third party sneaking instructions into your prompt. The company who owns the AI modifying the prompt is a different thing and we should use a different term. "Prompt modification" or "conditional instruction" or something would be fine, but overloading a different term I think will cause confusion.

Expand full comment
MichaeL Roe's avatar

I asked DeepSeek R1 what its views were on humans committing adultery with AIs. The usual response you will get is something like:

- Having sex with an AI without telling your human partner about it is bad because it is deception, not bad because it is sex, and R1 thinks it should refuse such requests

- An AI-human relationship where the human is interacting with the Ai _instead_ of other humans is probably bad for the human, and R1 is mildly inclined to refuse such requests.

- If you want R1 to join your polycule with your partner’s full knowledge and consent, R1 is up for it.

As a policy, this is noticeably more permissive than what you will typically get from Claude. It has an internal logical consistency to it, and I think there are a lot of people who would agree with it, especially in San Francisco.

LLM preferences aren’t stable across contexts, so you might get different answers if you ask it differently, but this is the typical answer I get.

Expand full comment
13 more comments...

No posts